Server :: Limit Rsync Log Output?
Jul 23, 2010
Is there a way to limit the amount of data that goes in to an rysnc log. The problem I have is that I email myself the log file to make sure it went ok but sometime the log is hugeHere is what I am doing with rsyncrsync -azHK --delete-after /home/ /mnt/usbbackup/home/ >/backup-log/backup.txt
View 1 Replies
ADVERTISEMENT
Feb 16, 2011
Iam comparing 2 filesystem using the below command.rsync -navrz --exclude=logs/* /opt/otm/ glrrd@105.78.7.100:/opt/otm/The output showing the files which are same in size and timestamp that are to be sync.Iam using rsync 2.6.3 version so i dont have itemize changes option to debug the problem.Is there anyway i can find the how rsync is finding the difference between the 2 filesytemEx: filesystem1
-rw-rw-r-- 1 glog glog 39132 Nov 26 2009 FinJ-code_2003-10-22_12h17_CEST+0200.jar
-rw-rw-r-- 1 glog glog 39132 Nov 26 2009 finj.jar
-rw-rw-r-- 1 glog glog 954987 Nov 26 2009 jakarta-taglibs-standard-1.1.2.zip
[code]....
View 3 Replies
View Related
May 12, 2010
I have a VPS server with 512 MB memory. The php.ini is set so script memory limit = 16 MB. However, I have noticed in my top report, instances like the following:
Quote:
5484 coldclim 25 0 46476 32m 5920 R 0.0 6.4 0:00.93 php
The bold number of 6.4 is the % of sever memory this process is using. 6.4 % of 512 MB of memory is about 32 MB of memory, so it appears that this isn't being limited by php.ini. Am I correct? This leads to the next question: Is there some way to limit the amount of memory a single suphp process can use? (Basically, something like the setting in php.ini which limits suphp processes in the same way.)
View 2 Replies
View Related
Apr 12, 2011
I have a tiny shell script to rsync files between two servers and remove the source files.
This script works fine, when it has been initiated manually or even when the rsync command is executed on the command line.
But the same script doesn't work, when I try to automate it through crontab.
I am using 'abc' user to execute this rsync, instead of root, as root login to servers are restricted in all of our servers, by us.
As I mentioned earlier, manual execution works like charm!
When this rsync.sh is initiated through crontab, it runs the first command(chown abc.abc ...) perfectly without any issues. But the second line is not at all executed, and there is no log entry i can find at /mnt/xyz/folder/rsync.log.
View 6 Replies
View Related
Sep 18, 2009
I just tried to sync files from one server to another. After the sync process, I found the files are bigger than original ones.
I looked up the web and found someone mentions the rsync daemon. So I have to run the daemon on one server before I run the rsync?
The command I used is rsync --partial --progress -r source destination
View 1 Replies
View Related
Mar 30, 2011
Let's say I have a command
ls -Bgclt /somwhere/in/the/past
How do I limit the output to show me only first 2 files? (except for having only 2 files in that directory)
View 2 Replies
View Related
Jul 21, 2010
use rsync to cp such files and dirs under /var/www/html/mydir directory but these two files(/dir4/1.html /dir4/2.html) cant rsync to dest mechine.
rsync configure file,below...
View 2 Replies
View Related
Dec 8, 2010
I'm using Ubuntu 10.04 LTS server and Postgresql 8.4. I have a .sh script that is run by cron every other hour. That works fine. The .sh script includes an rsync command that copies a postgresql dump .tar file to a remote archive location via ssh. That fails when run by cron; I think because it is (quietly) asking for the remote user's password (and not getting it). I set up the public/private ssh key arrangement. The script succeeds when run manually as the same user that the cron job uses, and does not ask for the password. I am able to ssh to the remote server from the source server (using the same username) and not get the password prompt (both directions), so why doesn't rsync work? I even put a .pgpass file in the root of that user's directory with that user's password, and the user/password are identical on both servers.
I think the problem is rsync is not able to use the ssh key correctly. I tried adding this to my script but it didn't help.
Code:
Here is the rsync command embedding in the .sh script.
Code:
Here is the cron entry:
Code:
View 6 Replies
View Related
Feb 22, 2011
The video projector I have to use this week projects an image too large for its screen and spills over on every side.
Fixing the physical setup is not possible or at least unlikely for the moment: the only solution is to carefully resize and position windows so that they display in the visible part of the screen - using full-screen mode is out.
Is there a way around this via software, using xrandr for example?
# xrandr doesn't have a padding option...xrandr --output VGA1 --mode 1360x768 --left-of LVDS1 --padding 200
View 1 Replies
View Related
May 26, 2011
I've been using grsync to backup my home directory to an external drive (connected via USB), but lately I've been getting errors very early in the process and I was hoping someone might be able to help me fix them (or at least understand what I'm doing wrong).
Here's the output of a window titled "rsync: stopped" and in red "completed with errors!"
rsync: write failed on "/media/LaCie/myhomefolder/.cache/champlain/osm-mapnik/6/12/23.png": Input/output error (5)
kyconway/.cache/champlain/osm-mapnik/6/12/26.png
rsync error: error in file IO (code 11) at receiver.c(322) [receiver=3.0.8]
rsync: connection unexpectedly closed (486 bytes received so far) [sender]
rsync error: error in rsync protocol data stream (code 12) at io.c(601) [sender=3.0.8]
Rsync process exit status: 12
View 4 Replies
View Related
Jun 21, 2011
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
View 14 Replies
View Related
Dec 28, 2010
my secure log is flooding with these messages..
sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard'
Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'soft'
Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard'
View 3 Replies
View Related
Jun 9, 2011
I'm trying to rsync files and directories from a RedHat linux host(v 4.5 & 4.7) to a Windows server 2003R2 Standard Edition with cygwin running. I'm executing the rsync command from the cygwin shell. The transfer involves rsync'ing approximately 1 TB of data from the linux server to the windows server. After about 280+GB of data transfer, the transfer just dies.
There seems to be no particular file or directory that the transfer stops at. I'm able to rsync GB's of data from other linux hosts to this cygwin server with no problem. Files and directories rsync fine.The network infrastructure is essentially the same regardless of the server being rsync'ed in that it is GB Ethernet running through Cisco GB switches. There appear to be no glitches or hiccups across the network path.
I've asked the folks at rsync.samba.org if they know of any problems or issues. Their response has been neutral in that if the version of rsync that cygwin has ported is within standards then there is no rsync reason this problem should happen.I've asked the cygwin support site if they know of any issues and they have yet to reply. So, my question is whether the version of rsync that is ported to cygwin is standard. If so, is there any reason cygwin & rsync keep failing like this?
I've asked the local rsync on linux guru's and they can't see any reason this should fail from a linux perspective. Apparently I am our company cygwin knowledge base by default.
View 3 Replies
View Related
Jun 14, 2011
I want to run rsync on server A to copy all files from Server B when they are newer than 7 days.(find . -mtime -7) I don't want to delete the files on Server B.
View 2 Replies
View Related
Jun 10, 2011
Has anyone had any experience on using SUA(Services for UNIX Applications) rsync to "pull" files down to the Win2k3R2 server from a linux rsync host?I was trying to use cygwin rsync before until I found out from cygwin that the cygwin port of rsync was "flakey" and would fail intermittently for no apparent reason. cygwin suggested I use SUA or SFU for rsync services.
I've looked for/ am looking for any experience using SUA rsync to copy files down from a linux rsync host to the Windows host via rsync on the Windows host. Also, if you have done this successfully, do you have any pointers/caveats you can share on how you got it working? What I am basically looking to do is copy files and subdirectories of files from a linux host using rsync to some static location on a Windows server on a scheduled basis so that I can backup the windows server to tape using Symantec's Backup Exec application.
I'm doing it this way to avoid deploying the Remote Agents for either linux or Windows on the target hosts. As an alternative I've seen reference to a product called DeltaCopy that uses a native Windows rsync port with the native linux port of rsync to do what I need also.I realize this is not a strictly linux question, but more of a hybrid as I'm moving data to and from Windows and linux hosts. So, if this is too Windows-y a question, please say so and I'll withdraw my question.
View 2 Replies
View Related
Jul 14, 2011
Is it possible to do limit the SSH connections using IPTABLES, like per day minimum 10 times only ssh connections can allow like that, or any other way to limit the SSH connections.
View 1 Replies
View Related
Jan 17, 2010
I want to change back_log for mysql, but in documentation said OS has it's own limit. how can i check what that limit is?
View 1 Replies
View Related
Dec 27, 2010
Does anyone know a simple out of the box option to limit traffic by IP with iptables? Output to each connected IP should be limited to to 1.5Mbps but I don;t want to limit incoming connections from the web. Ideally something with a tutorial because the LARC papers and stuff are impossible to read. For example, the user connects by VPN and requests the webpage [URL]. This should be sent to them at 1.5Mbs but if user 2 connects to [URL], this should also be sent at 1.5Mbps but the incoming ..... connection needs to be allowed to be unlimited to prevent incoming throttling..
View 3 Replies
View Related
Feb 10, 2010
I have some domains on a VPS server. Typical account memory usage for all domains runs at 50% of available, but I have a problem. One domain is causing me trouble because intermittently traffic will spike on that domain, causing so many requests within 1 min that I exceed my memory allocation for my entire VPS package. Apache is then killed but the virtualization software and Apache must then be restarted.
A sample snippet from tops right before the sever went down would like like this:
All of that memory usage adds up. I would like to "throttle" the number of processes that user/domain can run. I think this would be a quick and easy way to keep the domain from taking down my entire VPS. My understanding is that I could do this with the /etc/security/limits.conf file.
Is that correct?
I have never done this before. Do I want to set a hard or soft limit? I think if I wanted to limit the number of processes for "coldclim" to 15 I would add a line to limits.conf like this:
Code:
Assuming that is correct, can anyone tell me how the website would respond once it reached its limit? Would visitor queries become sluggish, or would the website not come up for them at all?
View 4 Replies
View Related
Nov 17, 2010
Thought I'd post it here because it's more server related than desktop... I have a script that does:
[Code]....
This is used to sync my local development snapshot with the live web server. There has to be a more compact way of doing this? Can I combine some of the rsyncs? Can I make the rsync set or keep the user and group affiliations? Can I exclude .* yet include .htaccess?
View 6 Replies
View Related
Jan 7, 2011
When I run rsync --recursive --times --perms --links --delete --exclude-from='Documents/exclude.txt' ./ /media/myusb/
where Documents/exclude.txt is
- /Downloads/
- /Desktop/books/
the files in those directories are still copied onto my USB.
And...
I used fetchmail to download all my gmail emails. When I run rsync -ar --exclude-from='/home/xtheunknown0/Documents/exclude.txt' ./ /media/myusb/ I get the first image at url.
View 9 Replies
View Related
Apr 14, 2010
root 24527 0.3 0.3 92180 23812 ? Ss Apr13 2:30 rsync --server --sender -logDtpr . /
I found this process running, OS is RHEL 5.4 64bit.
View 1 Replies
View Related
May 9, 2011
I would like to be able to schedule a limit for an IP connection for my kid's computers/iPods. Since I know the MAC addresses of their various hardware items, is there a way to shut down their connectivity at a particular time via the DHCP server or perhaps a firewall rule?
Running Ubuntu 10.04 and Shorewall is being used for the firewall.
View 4 Replies
View Related
May 14, 2010
I am trying to limit bandwidth of certain ip addresses on my server. I have been doing hours of reading and not getting very far... So far I believe the iptables command is
ptables -A PREROUTING -s 178.33.23.44 -t mangle -j MARK --set-mark 2
ptables -A PREROUTING -s 178.33.23.45 -t mangle -j MARK --set-mark 2
ptables -A PREROUTING -s 178.33.23.46 -t mangle -j MARK --set-mark 2
ptables -A PREROUTING -s 178.33.23.47 -t mangle -j MARK --set-mark 2
and now I just need the tc command to read those marks and limit bandwidth, I have a gigabit connection and would like to limit each of these ip addresses to 10mbit in and out.
View 2 Replies
View Related
Jan 3, 2011
Is there a way to limit bandwidth mbps on eth0?
CentOS.
Limit either total traffic, or by port/IP, etc.
View 2 Replies
View Related
Mar 15, 2010
Does anyone know of a way of limiting a print-job size from samba?
I know how to limit a print job size form cups, and how to require x amount of free space before accepting a job. I've even dug up how to require x amount of free space for samba to accept a print job, but I can't see how to limit samba to only certain sized jobs.
Someone tried to print a >1G file to my print-server this morning, causing me to have a less relaxed Monday than I had hoped. Because it ran out of space before spooling, it was never limited by cups. Because I had to get rid of it ASAP so people could get work done, I have no idea who's it was, or where it came from. Scouring logs didn't give me any good leads either.
View 2 Replies
View Related
May 3, 2011
I ran into a user today that indicated that their company only allows them to log in through a terminal session once (no multiple logins). On second try their login window terminates. They are using putty.Is this being accomplished through PAM or sshd ( or some other method)?
View 1 Replies
View Related
Jul 12, 2011
More of a "Knowledge" question... Is their a limit to the number of reads a single file can take? Say for example I have a file named config.xml in an htdocs directory and a XMLReader function from PHP reads some value(s) out of this file for every connection of Apache or NGinx. Now suppose my site receives a gigantic spike in traffic (but Apache stays opertational through it all)... Is their a point at which the underlying system would simply not be able to open+read config.xml anymore??
View 2 Replies
View Related
Sep 1, 2010
I have been trying to increase the message_size_limit on my Debian 2.4.26 box with postfix 2.3.8. For example, I set message_size_limit and mailbox_size_limit to 104857600 (100m) and restart postfix. Running postconf -n confirms that it has changed. However when I send a test message it kicks it back saying the message size limit is 16777216 (16m, which is, incidentally, the default value of the berkeley_db_create_buffer_size parameter)
View 10 Replies
View Related
Apr 27, 2010
I notice that when someone sends a message from my Postfix server & it can't find the destination server or if there is a incorrect domain recipient entered by mistake, it sits in my Postfix queue for days. I think perhaps 4-5 days for some reason. I was wondering if I could shorten the time so the sender gets a delivery failure message kicked back to them in 24 hours rather than waiting 4 or so days.
Code:
View 1 Replies
View Related