CentOS 5 :: Time Changes In Rsync Log File
Mar 18, 2010I'm running rsync in daemon mode on CentOS 5.4, and I'm seeing inconsistent times reported in the log file.
View 2 RepliesI'm running rsync in daemon mode on CentOS 5.4, and I'm seeing inconsistent times reported in the log file.
View 2 RepliesIm looking for an app pr line of code that could let me observe a process, save the info in a number of variable and then put the gathered info on a file.
Ive been trying with variations of top but no luck. I am running several CentOS virtual servers, VM is 2gb ram 2 processor.
Maybe a script that works over a specified amount of time while writing lines with the info on a text file so at the end i can have a sort of table with the data.
The thing is Im going to stress test the server and I would like to have the data to make some statistics.
this idea to me right now seems stupid and impractical so I'm also looking for solutions outside of this one. I'm creating a set of highly available servers. They currently utilize pacemaker to manage a shared IP address, apache management, mysql management and also drbd handling. They currently replicate, properly failover and run fine. However, this is a shared web hosting setup and everytime a new user is added to the system a username and password is created and the httpd.conf file is updated for their site. I need a way to replicate those changes to the other server. For httpd.conf I'm pretty sure I can just stick it on the apache drbd resource and create a symbolic link. However, for the passwd and shadow files that is not a realistic solution.
View 1 Replies View RelatedI have a CentOS5 server with a 1tb hard drive.There is only 80gb of data on that huge drive and now I want to make a bare metal recovery backup using AcronisMy question is, how can I estimate the amount of time the backup will take and the size of the image file? Is it based on the size of my drive or is it based on the amount of data on the drive?
View 1 Replies View RelatedI have two servers, both running CentOS, but one running Cpanel as the control panel and the other running Webmin.
I would like to sync the web folders using rsync. However, when I sync the files from the Cpanel host the permissions look as follows:
Code:
But that doesn't work on the Webmin server which needs the following:
Code:
It seems the Cpanel install has the group as 'nobody' whilst webmin has it the same as the file owner. I can fix this by running:
Code:
But as the file transfers will be frequent, I don't want the files to be inaccessible until chown is run.
So, what I really want to know is - how can I change things on the 'webmin' server so that files with ownership owner:nobody will run?
I have a Linux box, running rsync, only problem is, I have to turn it off at night. What I want to do is turn it on and when it turns on, rsync will already be running in daemon mode rather then having to run rsync daemon. Does anyone know the solution? I'm sure its simple, I just cannot figure it out
View 3 Replies View RelatedIn the right upper corner of your screen,you can see if you have a network connection or not.Most of the time I do not have one according to Networkmanager (?).So I have to enable my wired connection manually.I already tried a lot to change my settings so I would have a network connection at boot,but it doesn't seem to work....Firefox is in Offline modus when I logon..
View 2 Replies View RelatedI am hosting two Virtual Servers both running Centos 5.3 on a host machine also running the same OS. The VM software in use is Xen, as supplied with the OS.The host machine's time and date is fine, however both Virtual Servers are running ahead of real time consitantly.Running /etc/init.d/ntpd restart will resolve the issue however one of these is running MailScanner and when the time suddenly goes backwards, sometimes by as much as an hour, it stops working properly.
View 1 Replies View RelatedI installed CentOS 5.4 on a virtual server to host the tool OpenVAS, which is used to execute vulnerability assessment.This tool is based on a master engine and thousand of plugins, each one doing a specific, little function.I wrote a cron that, every sunday, updates the list of these plugins to keep the system updated; this cron uses a tool's proprietary function that download the new plugins from a site using rsync.I was using the rsync 2.6.8-3.1 until last week, when I decided to install ALL the updates proposed by operating system itself (more than 170) and the whole list includes rsync 3.0.7-1.el5.rf (from rpmforge).
This monday, the log of the tool's function gave me a bad response about the impossibility to establish a connection with the remote server due to a connection timeout.Before doing the rollback of the system, I tried to understand what happened, so I started manually the function on a terminal and, together, I was watching the results using netstat -tuvnap, so I saw that there was an attempt of connection, between my server and the remote one, but the connection did not finish the three way handshake, because the state of connection was always "SYN_STAT".I tried to value the $RSYNC_PROXY variable, because there is a proxy between my server and the world, but the result was exactly the same, so I was forced to execute the rollback of the system, which works very well now!
I'm making my own yum repository - firstly so that all the machines I administer can be updatedvia the internal network, secondly so that I can test any updates on a spare machine before passing them on, and thirdly so that I can add my own repo for internal software.
I've created the necessary folders under my webserver, and used rsync to update them from my local CentOS mirror, following the instructions at [URL]
I notice it says to run "createrepo" on the base repository, created by copying the rpms from release DVD.
When I rsynced the updates repo, but I notice that the files in repodata are very small. In fact, having a look inside them, filelists.xml contains no file details. But if I run createrepo in the updates directory, filelists.xml gets lots of file details inside it.
I wondered if maybe the local mirror hadn't been updated properly, but checking against mirror.centos.org shows that has the same files.
how does the (real, live, CentOS) updates repo work when there is nothing in the filelists?
I have a script that users execute to copy files from one server to another 3 boxes.
I want to prevent rsync from trying to copy files with spaces.
Anyone have an example I can use?
The rsync just craps out when they try to sync files with spaces, I tell them that but....
Is there a graphical interface for centos 5.3
View 3 Replies View RelatedI believe it was rsync was the tool. I have a box running CentOS 5. It has a 250GB HD in it. I have another drive with the exact model. Currently it has a as a said a 250GB IDE drive. I want to shutdown the machine, install this other hard drive and set up a cron job that will backup my main drive at times that I set. This way if the main drive fails, I do not loose all the data and have to rebuild the server from scratch as I have been custom configuring it for years. I can't remember if there was an issue with the main drive being mounted to back it up or not. I have looked at some of the how's on rsync but they seem to only talk about using another server for this. If I shit down the box, install the new drive, and the box boots back up, is it going to ask about that drive or what do I need to do to get rsync going and does it partition the drive as such? And can I do it this way. This way if the main drive fails, I can just swap the drive and be on my marry way.
View 2 Replies View RelatedI want to use rsync to backup a directory on my local hard drive to a USB hard drive.
I created the initial backup with:
rsync -av /share/chris/docs/* /usb/docs
This command would copy the files to the local directory,find /mnt/nas -type f -ctime 1 -iname '*.avi' -exec rsync -av {} /mnt/Mythbuntu
View 1 Replies View RelatedI am trying to setup a Rsync backup method.I have multiple files and folders all OWNED by different users.Now I can read the files as root.But I want to make a user for RSYNC that can ONLY read all these files for copying them across.It requires to create a non password login and I dont want to do that with the root user ! I have tried to create a new user and use 'setfacl':setfacl -m u:someuser:r /var/test.But when I su to 'someuser' I cannot copy the files to my homedir. So I think it wont work for rsync also.
View 8 Replies View RelatedNeed to know if there are any known errors allocating disk space using rsync. I have recently setup a small file backup server (2 500GB drives) running Centos 5.2, with a rsync daemon (via xinetd). However I have noticed that after about a week of usage, if I reboot, then the disk that rsync uses, has a lot of multiply claimed blocks and inode errors, such that I have to run fsck multiple times to fix. This has happened a couple of times now so I don't think that its random.
View 11 Replies View RelatedI am running version 5.5. I am using Grsync to back up 2 folders to an USB external 1.5 terabyte drive. Sometimes grsync backs the folders up correctly but most of the time, I need to unmount and then mount the drive to get it to work.
View 5 Replies View RelatedI'm using rsync to create a mirror of the data files on our main server every day. I've looked at the man page, and can't see it; can I get a listing of the files that have been changed on or added to the mirror when it's completed? Can it just log what it's doing to a file?
View 1 Replies View Relatedhow rsync will handle disk images. Will rsync copy only the changed blocks of a vhd or a lun? This is what I've been told, but wouldn't this require overlaying a filesystem on the vhd? How would rsync handle copying a 500GB lun?
View 1 Replies View Relatedi am trying to transfer a file from my live linux machine to remote linux machine it is a mail server and single .tar.gz file include all data. but during transfer it stop working. how can i work and trouble shooot the matter. is there any better way then this to transfer huge 14 gb file over network,vpn,wan transfer. the speed is 1mbps,rest of the file it copy it.
rsync -avz --stats bkup_1.tar.gz root@10.1.1.22:/var/opt/bkup
[root@sa1 logs_os_backup]# less remote.log
Wed Mar 10 09:12:01 AST 2010
building file list ... done
bkup_1.tar.gz
deflate on token returned 0 (87164 bytes left)
rsync error: error in rsync protocol data stream (code 12) at token.c(274)
building file list ... done
code....
I have 2 different mounts. One points to a local windows share(NTFS ->Samba) and the other one points to a PPTP VPN connection sharing(I belive that is NTFS too). I use "cifs" scheme in my fstab to mount these. And I use my Debian box to copy between these 2 mounts. I have started using Rsync for that purpose, I think that it works fine for now. My main problem is that it looks like Rsync cannot figure out if the files are same or not in source and target folders when I use these mounts. Most of the time Rsync copies the same files and folders over and over again even though those files and folders are on the target.
I am wondering if there is a way to make this scheme work? Being on a Vpn connection(slow) a Windows box, Rsync could have save a lot of my time if it could have recognized the files and folders that are same on both ends
I have successfully backed up my files using a script to a remote server with a log file output.However the log file is appended each time.I wish to have a different log file each time with date and time and have yet to figure that part out.
View 2 Replies View RelatedThought I'd post it here because it's more server related than desktop... I have a script that does:
[Code]....
This is used to sync my local development snapshot with the live web server. There has to be a more compact way of doing this? Can I combine some of the rsyncs? Can I make the rsync set or keep the user and group affiliations? Can I exclude .* yet include .htaccess?
When I run rsync --recursive --times --perms --links --delete --exclude-from='Documents/exclude.txt' ./ /media/myusb/
where Documents/exclude.txt is
- /Downloads/
- /Desktop/books/
the files in those directories are still copied onto my USB.
And...
I used fetchmail to download all my gmail emails. When I run rsync -ar --exclude-from='/home/xtheunknown0/Documents/exclude.txt' ./ /media/myusb/ I get the first image at url.
I have a tiny shell script to rsync files between two servers and remove the source files.
This script works fine, when it has been initiated manually or even when the rsync command is executed on the command line.
But the same script doesn't work, when I try to automate it through crontab.
I am using 'abc' user to execute this rsync, instead of root, as root login to servers are restricted in all of our servers, by us.
As I mentioned earlier, manual execution works like charm!
When this rsync.sh is initiated through crontab, it runs the first command(chown abc.abc ...) perfectly without any issues. But the second line is not at all executed, and there is no log entry i can find at /mnt/xyz/folder/rsync.log.
I just tried to sync files from one server to another. After the sync process, I found the files are bigger than original ones.
I looked up the web and found someone mentions the rsync daemon. So I have to run the daemon on one server before I run the rsync?
The command I used is rsync --partial --progress -r source destination
I made a shell backup script that uses Rsync and I am trying to get rid of the password prompt because it will use a CRON to run. I have set my variable in my shell script at:
PASSWORD_FILE=rsync_password
And the password in that file only takes up 1 line.
However when I run (ignore $DESTINATION)
rsync -aRvz tmp $DESTINATION --password-file=$PASSWORD_FILE
It still gives me the prompt. How can I accomplish this? I cannot allow a prompt and I do not want to have to use keys.
I'm always getting this error when I run the rsync command:
Code:
when I try to navigate the file:
Code:
it gives
Code:
I have a backup sh file that I have been using for a long time. It has always worked. 2 Days back I switched to a different pc and now suddenly the script don't work.If I run it manually in the terminal it works. But when it execute with cron it doesn't copy any files to the backup destination. It starts but doesn't copy anything.Can someone help me as to why it works manually but not with cron ?
View 9 Replies View Related