I have a linux box running between my router and my LAN. My connection speed is 10MB download and 1MB upload. The issue is that whenever someone starts to upload something, it is like my connection is down. No one else can open websites, read emails etc.Is it possible to place a limit for upload, maybe 50kb/s? This way, people won't use the entire upload speed available.
I'm a newbie in the world of netfilter/iptables. I've read an article about iptables and rate limit module: Code: iptables -A INPUT -p ICMP --icmp-type echo-request -m limit --limit 1/minute --limit-burst 5 -j ACCEPT The firewall will let the first 5 packets in in the first minute, -limit-burst 5; this means, however, that the packets/minute now is 5, so any further packets are blocked until packets/minute = 1, i.e. 5 minutes later. In the sixth minute, packets/minute will be 5/6 < 1, so another ping request will be let in. When the extra ping request is admitted, the ratio becomes 6/6 = 1 again, and packets are DROPped again until the next minute.
Now I have some problems in understanding how it works. For example: I want ping google.com in this way: the kernel firewall permits to send the first 5 packet to google.com (--limit-burst 5) and then it blocks the remaining packets for 5 minutes. At sixth minute (because I wish a limit rate equal to 1/minute: --limit 1/minute) one packet can send to google again. And so on.
So my rule should be: Code: iptables -A OUTPUT -d url_of_google -p icmp --icmp-type echo-request -m limit --limit 1/minute --limit-burst 5 -j ACCEPT In this way, if i digit Code: ping -f url_of_gogle I expect that the first 5 packets are accepted (and so zero '.' will print on the screen) and then for the remaining 5 minutes no one packets will be accepted (and so a long string of '.' will print). But it doesn't work...
In man pages of ping we read (about -f option): -f Flood ping. Outputs packets as fast as they come back or one hundred times per second, whichever is more. For every ECHO_REQUEST sent a period ``.'' is printed, while for every ECHO_REPLY received a backspace is printed. This provides a rapid display of how many packets are being dropped.
I have a large file (deflated size: 602191947)that is not saved in my Ubuntu One account. On sync'ing the file is being uploaded, and eventually reaches 602191947 - and then nothing more happens to this file - but sync'ing the following files in the queue goes on with success. I have tried manual upload with the same result. The file is still being marked as 'uploading' even after several tries and log ins/log outs, and reboots. So I was just wondering whether there is a file size limit - can't seem to find information regarding this.
There is a ps3 in my house which i play with an online connection, i also download alot of stuff. so i want to limit the upload speed of the ps3 to give my downloads more speed. i have verizon fios.
I use vsftpd. I would like to know if there is a way to limit the upload speed. Only the upload speed so the "anon_max_rate" and "local_max_rate" don't do the job. Someone told me about the trickle but I expected a more recent application.
I have a SSH server set up at home listening on port 22. I have hardened the server so it is pretty secure but I want to make it even safer by editing my iptables to rate-limit incoming connections and DROP false login attempts. I have tried these tutorials but I just cant get it to work:[URL]I want the debian-administration.org tutorial to work but when I try to add the first rule in terminal:sudo iptables -I INPUT -p tcp --dport 22 -i eth0 -m state --state NEW -m recent --setI get the following:Bad argument --set'I am new to iptables and I'm not sure if I'm doing something wrong when I try to set it up. I'm using Ubuntu 10.04.1 LTS with iptables v1.4.4.
I have already tried trickle and wondershaper. I need a program that can limit the speed of download/upload of an already running program. Similar to the how NetLimiter in Windows limits already running processes. Using Linux.
I just want to place upload/download limit to each ip address ( say 20MB per day for each ip or specific ip) using squid proxy i tried through delay pool but it control only download/upload speed not place any download/upload quota restriction. following are my codes I have 1Mbps line
delay_pools 2 delay_class 1 2 # here 700kbps(87Kbytes) Net-Total usage limit, with 50Kbytes per user
[code]...
I also used
"reply_body_max_size" parameter to control upload limit.
But
"request_body_max_size" only restrict download based on size.
how to setup quota ( download quota per day per ip)
i have VPS server and i installed Xserver on it and all ok i created new user for my client but i need to limit his access to the following
he can download and upload to his home file " browser by Firefox" he can't install or use any application "just the one i installed it" he can't see the file system or browser it !! if i can give him specific space on harddisk would be better he can extract and compress files he can't edit the settings ....
i have another sensitive folder and setting i don't want him to see it so how to limit his access?
have setup a LAMP server with ubuntu 10.04 server edition x86 for my study in VMware Workstation 7.1 For a assignment I had to make a php script that would load a file up to the server and set the name in a mysql database. According to the book the server should set the image in a cryptic folder in the /tmp/ folder.
This isn't working and i also try'd locate and find to find the image i uploaded. I checked the php.ini and file uploads were on but no folder so i set that one to /tmp/ but still no images. Can anyone help me with enabling this function?
sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'soft' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard'
I have a VPS server with 512 MB memory. The php.ini is set so script memory limit = 16 MB. However, I have noticed in my top report, instances like the following:
The bold number of 6.4 is the % of sever memory this process is using. 6.4 % of 512 MB of memory is about 32 MB of memory, so it appears that this isn't being limited by php.ini. Am I correct? This leads to the next question: Is there some way to limit the amount of memory a single suphp process can use? (Basically, something like the setting in php.ini which limits suphp processes in the same way.)
I have a LaCie NAS which is mounted on my main linux machine over a wifi LAN using the cifs file system. I would stupidly expect the transfer rate between my hard drive and the NAS to be limited by the Wifi speed (54 Mbps) but when I transfer files, the speed tops at 1.9 Mb/s which is roughly 15.2 Mbps. The most puzzling thing is that when I do multiple simultaneous transfers, I reach approximately 3MB/s in total but none of the individual transfers goes beyond 1.8. Does anyone have an idea about what is keeping the transfer rate so low?
Can anybody tell me what is rate control and rate control mode of operation (in data link layer) in an Ethernet interface?. Does it have anything to do with auto negotiation in Ethernet?
I want to implement rate control over network interface. So I have configured Linux PC as a router with netem installed and having two interface cards.
device1----> Linux router with netem ---> device2
device1 connects to eth0 and device2 connects to eth1 of the Linux router. eth0 is configured to connect to the internet and eth1 has a static IP address on a local network. I want to limit bandwidth on devices connected to eth0.So I applied the below rules using tc and tbf.
We have a DSL , how can we calculate the available bandwidth so that we can shape the packet, We can assume the bandwidth to be 100mbps on the ethernet interface However, in DSL devices, the train rate(Bandwidth rate ) is varies according to different situation. How can i get the available bandwidth rate(from varies)?. Any method is there for getting the available bandwidth value....
I'm sure this is simple but I haven't a clue how to do it other than System Monitor, which only displays for the total uptime.
I want to see what my total network data use is in Bytes (GB, MB, or KB) downloaded/uploaded for the month or any given date range.
Rationale is I want to use wireless tethering from my phone to replace my broadband connection at home. At the moment my phone contract gives me unlimited data access, however if I were to sign a new contract today this would drop to 2 GB a month with hefty overage charges.
We have a DSL , how can we calculate the available bandwidth so that we can shape the packet, We can assume the bandwidth to be 100mbps on the ethernet interface However, in DSL devices, the train rate(Bandwidth rate ) is varies according to different situation. How can i get the available bandwidth rate(from varies)?. Any method is there for getting the available bandwidth value..
We have a Linux box which acts a a file server. Currently, files and directories are exported using NFS.At the moment, we are a bit concern on its data transfer performance. FYI, we are using a embedded Gigabit Ethernet port on the file server. We ran a few simple write tests between NFS client (also utilizes GigE port) and the NFS server. In these tests, both NFS server and client are both connected directly to each other with a Cat5E cable. Unfortunately, the write/transfer speed results are not as per our expectation. It scores roughly about 11-12MByte/s, where as theoretically Gigabit Ethernet transfer rate is able to reach up to approximately 120MByte/s.I wouldn't expect to reach the theoretical max transfer rate (it would be great if we can , but I would appreciate if you guys can share with us in terms of the following :
1) What's the practical max data transfer rate which you guys managed to observe in a normal Gigabit based connection? What about jumbo frames configuration?
2) Is there any additional tuning/configuration we need to do within the OS to reach those practical max data transfer rate figure?
3) Does PCI-e / system bus plays a role in achieving this speed? For example, we are using the embedded GigE port and we heard some people says embedded ports are actually sharing the system bus and resources with other devices, which might adds into performance issue. Correct me if I'm wrong.
4) Does converting to Cat6 cabling will guarantee an increase in the data transfer performance?
5) In the future (once we are clear on how much single GigE transfer rate we can go) , we are looking into doing bonding since that the NFS server's shared directory/volume read-write speed is way much higher (i.e 400-600MByte/s). Will bonding allow us to achieve higher NFS read/write speed? What are the bonding modes best used for this purposes? Appreciate if anybody who has experience in doing bonding for NFS can share their experience.
I have a FC13 box that has both Gnome and KDE sessions installed.
I have noticed on the KDE session that data transfer rates are slower than when I use Gnome.
In Gnome, I can transfer files between my FC13 machine and my Ubuntu 10.04 pc at a rate of 6.5 MB/s (52 Mb/s if my maths is correct), but in KDE the rate is only 3.5 MB/s (28 Mb/s).
"ethtool eth0" shows my NIC speed as 100 Mb/s. Obviously I am not hitting anywhere near that speed in either session, (a separate article may be happen in the future to address that), but I am curious as to why KDE is that much slower for file transfer.
The wireless usb adapter connects to my network, and the router is running on mixed mode, except that I cannot seem to get wireless 'n' connection speeds. Both are capable of wireless n. I check the connection info and it is always 54 Mbps. I have set the router wireless on N-only and I lose my connection.Is the standard driver for this usb adapter in Ubuntu 10.4 unable to utilize 'n' speeds?
left to its own devices my wireless connects at too low a speed. I have a 20meg internet connection and my wireless is slowing it down to like 3meg. When I reboot into windows it's fine. When I run iwconfig eth1 rate 11M or even 24M the connection is much faster and runs fine, why won't it automatically go higher? Is this the fault of the driver? I am running Broadcom's driver compiled from source. Also, how can I set it to 24M at boot?
Edit: I guess I can put iwconfig eth1 rate 24M in rc.local, but I'd still like to get auto running higher, especially as my signal should be pretty good. Output from iwconfig white rate=auto:
eth1 IEEE 802.11 ESSID:"honeypot" Mode:Managed Frequency:2.417 GHz Access Point: xxx Bit Rate=1 Mb/s Tx-Power:24 dBm Retry min limit:7 RTS thrff Fragment thrff Encryption keyff
I am struggling trying to understand the reason for a fairly slow data transfer rate between two machines. ( tried point to point and also via a 1 gb switch ) One is nfs/http/ftp server ( with raid1 and lvm on top ), the other one my desktop pc. Both OS with default options, no changes to kernel in proc or other sort of thing.
Hardware is full recognized and perfectly working: The server has 4gb ram, Intel Core 2 Duo CPUE6850 @ 3.00GHz, 1000Mb/s NIC card and Lucid 10.04 64 bit, 250Giga Hard disk. The client has 3gb ram, Intel Core 2 CPU 6320 @ 1.86GHz, 1000Mb/s NIC card and Ubuntu Maverick 32bit , 150Gb Hard disk.
Raw data is good: gettons@gettons-desktop:~$ iperf -c MYSERVER ------------------------------------------------------------ Client connecting to MYSERVER, TCP port 5001
I have not tried this, and I am only wondering about the result.Let's say that I have a PC/Laptop with two network devices: an ethernet and a wireless. Can I connect both of the to the same network (if this network allows both connection) to increase the transfer rate between the PC/Laptop and the server???
I have a slave node uploading all kinds of backups to my server in the internet. Now I would like to display the actual upload and download rate to this server (not the entire nic-traffic, any protocol) in a small php-page for easy monitoring.I had a look at quite some monitoring tools and the one which kind of offers what I am looking for is iftop with a filter on the IP of my server. As I would like to periodically update a file with the actual rates, an interactive program won't do. A possibility would be to filter the packages myself using but this seems to be quite a long shot.The optimal solution would be a program or script printing out the actual upload to a host specified in the options to STDOUT