Ubuntu Networking :: Limit Amount Of Data To Be Received?
Jun 30, 2010
I was wondering if there was a Windows or Ubuntu way to limit the amount of data that is able to be sent over the internet between certain times, eg. Between the times of 7am and 7pm can only download 300 MB from the web, when this limit is reached the web is either disconnected or slowed down.
I want to use the tc rules for bandwidth control in my lan.i have a linux router(traffic shaper).how i can limit the amount of bandwidth every user has access to per day? For example, any user can't download more than 2 gb per day(or per month).
The information about lan's users(such as a group type,userid,etc.but not any thing about time limiting per user)is in ldap directory on ldap server.the linux server uses ldap server for authentication users when the user login.
I have a Webserver (Co-Location) and all runs fine ... since last week. Now there are a lot of RX-ERR shown in netstat and ifconfig. And when I try to upload a external website direct on the server for example via wget, it is very very slow and hangs very often.
I have analyse the network but I was not able to find a problem. My hoster has checked the network and all looks fine. For example my hoster has plugged-in a pc in the same switch ... and was able to do wget (load external data, like websites) in normal speed.
Since last week my websites were delivered slower as before, too. It seemed there is a network-problem ... but how can I find it?
Actually I can install moduls ... but the server needs hours. So, if you knows a good command-line tool to analyse the network.
my isp is putting a max bandwidth in my area and I need to monitor my downloads and uploads per month. Is there anything that has a gui that is easy to set up and just shows the amount of data downloaded and uploaded per month. Also if possible to do a pop up if you set a maximum bandwidth amount.
When I run ardour sound editing I get this message , but it starts ok Your system has a limit for maximum amount of locked memory! This might cause Ardour to run out of memory before your system runs out of memory. You can view the memory limit with 'ulimit -l', and it is normally controlled by /etc/security/limits.conf
bash-4.1$ ulimit -l 64 my limits.conf is like this audio - rtprio 99 @audio - memlock 250000
I have a few multi-user servers in an academic laboratory. I am having a problem with some users maxing out the available RAM, causing such sever slowdowns the machine essentially crashes. My servers are Dell Power Edge's running Ubuntu 8.10 Server Edition (Not my choice). I would like to set a maximum limit on the amount of ram a user can utilize. This morning I experimented with setting limits via /etc/security/limits.conf and using ulimit. Neither of them prevented my test program, a simple infinite loop of mallocs, from crashing the server.
i'm trying to write a program with c socket programming,what i am trying to reach is a program which will calculate a computer's downloaded data from the internet,just to know how much he/she download?
My machine has 4 SATA 2 West Digital 1TB disks. I made 2 partitions on each of them, 500GB for each partition. When I started using them I check their I/O using iozone. The first partition has 100MB/s for read, 70MB/s for write. And the second partition has 80MB/s for read, 55MB/s for write. All 4 disks has the same result.
As I use on, the I/O speed on each partition decrease, to different extend. For example, for the 4 first partitions, the write speed varies from 69MB/s to 56MB/s. And I have same amount of data on each of them, all used 11%.
My guess for this is the disk block allocation policy. This is caused because some disk starts writing from inner location while others writes on the outer edge, even though amount of data on each disk is the same.
I have a probably kind of unusual problem - when a USB stick is connected to the PC and data is copied from/to that stick, I need to know how much data has been copied. The data itself if not interesting, just how many bytes. I unfortunately don't have access to the program that does the copying, and most of the data doesn't end up on any drive (it just gets read and discarded), so I can't simply check the size of a target directory or something like that. I have had a look at usbmon, but that seems to produce way too much data - the normal case would be around 10 gig of data being read, and I can't have that blown up by a factor of 10 and lying around on the hard drive
Elsewhere is my question about 11.3. I have had to abandon that and 11.3 as my RAM is too small. Now, for the first time in many years, I find it necessary to extract specific data from received .pdf files. According to OpenOffice, editing of these files is only possible from version 3.2. My concern is whether this later vesion will be compatible with 11.2.OpenOffice installed is 126.96.36.199-1.1.4-i586. The same question has been directed to their Forum.
My DVD player attached to my TV is picky. Burn CD's (with avi files) at 8X and it plays them. 16X and it complains about bad disk. Every time. In karmic, I can only get cdrecord to burn at 16x.
Here's the command: cdrecord -dev=/dev/scd0 -speed=8 -dao -pad -v something.iso
And it burns at 16x instead of 8x. Command issued with root privilege. Identical hardware under CentOS 5.x and it burns at 8x, like I asked it to. I'm all SATA if it matters. What do I need to do to get the version of cdrecord with karmic to burn a data CD at 8x? For whatever reason, CDs burned at 16x just don't work in the DVD player attached to my TV.
I am new to Ubuntu almost installed it after windows showed blue screen 4 ever n ever. However after installing ubuntu whenever I log in windows it doesnt detect any network connection but when I use Ubuntu it automatically does can it be that Ubuntu is causing any problem? And how to check how many packets are being sent and received through my wired network and do I have to install any drivers for my modem in Ubuntu.
This has been bugging me for quite a long time. I need to have a pptp connection to my university. I have tried to locate the error, but this has been a real trouble.
A little bit of history: My girlfriend had a laptop with M$ vista on it with a working PPTP connection. I had a laptop with Ubuntu, with which I couldn't get the PPTP working.
When my laptop crashed, I bought a new one on which I run a dual boot XP SP3 and Ubuntu. I also own a desktop (Dual M$ Win 7 / Ubuntu standard Karmic) and my girfriend also upgraded to Win 7.
Currently, the XP version is the only one on which I can get a connection. Both Ubuntu and the Win 7's give me a GRE proc 47 not received error. I know our router (a DAVOLINK DV2020) is notorious for not sending through GRE proc 47, but this either shouldn't be the case or it shouldn't matter, since I can make a perfect connection through the XP OS.
This is quite a handicap, since we both study at the university for which we require working VPN PPTP connections.
I am having (seemingly) random trouble with my wired network ever since I installed Lucid. I have no problem getting an ip address from dhcp. However, randomly the computer will boot and although I have an ip address I do not receive any responses for pings on the network nor can I browse the web. If I sudo /etc/init.d/networking restart a few times (or reboot) it will start working. However, restarting the networking services (as mentioned above) again will cause me to no longer receive responses for pings or browse the web.
Furthermore, I have never been able to successfully ping if I manually set an ip address. I have un-installed network manager and I am using /etc/network/interfaces to configure the network. Using Lucid Lynx 64bit on a Dell Precision. I have pasted below the output of a few working commands. When I switch between static ip and dhcp I am commenting/uncommenting the lines shown in /etc/network/interfaces.
I set up environment according to BareMetalProvisioningBestPractices document, I think everything should be ok. Now trying to plug in new virtual machine for PXE install Linux. But something is wrong, DHCP waits couple of seconds, and then I receive:
PXE-E53: No boot filename received PXE-M0F: Exiting Intel PXE ROM.
I think, all services required are started.
dhcp.conf on boot/stage is: # # DHCP Server Configuration file. # see /usr/share/doc/dhcp*/dhcpd.conf.sample #
pxelinux.0 was taken from RHEL4, but I've got image and trying to boot RHEL5. Maybe there is something?
I used to connect to the internet via a Linksys router using the following terminal commands (Network Manager has never worked for me on any Ubuntu install):
sudo ifconfig wlan0 down sudo dhclient -r wlan0 sudo ifconfig wlan0 up
However, I recently changed the old modem + router for a new wireless modem (a Thomson TG585v8 ). Now, the result of that last command is this:
amosupremo@amosupremo:~$ sudo dhclient wlan0 Internet Systems Consortium DHCP Client V3.1.3
No working leases in persistent database-sleeping. I've tried to connect in Natty(both Ubuntu and Kubuntu) and it's not working either.I also purged Network Manager and installed wicd and I managed to get an intermittent, slow connection.
1 ) Machine Brand and Model (PC/Laptop): No brand. I built it with the following specs:
AMD Athlon II X4 2.6GHz 2 HD: 100GB Sata (Ubuntu and XP64 install) / 80GB IDE (XP) 4GB RAM Gigabyte Motherboard
2 ) Wireless Brand, Model and Wireless Chipset:
amosupremo@amosupremo:~$ lsusb Bus 002 Device 005: ID 0ace:1201 ZyDAS 802.11b Bus 002 Device 004: ID 049f:000e Compaq Computer Corp. Internet Keyboard
3 ) check interface:
amosupremo@amosupremo:~$ ifconfig wlan0 wlan0 Link encap:Ethernet HWaddr 00:02:72:04:d0:04
I am also attaching the wicd log. It contains a session where I got the intermittent connection. I stopped that connection and restarted it (with the same results) two times.
i am using centos 5.2 . recently i have implementing transparent proxy but i faced issue. i received lots of messages in dmesg. i have 2 machine with same OS another machine are works fine with this same configuration
br0: received tcn bpdu on port 1(eth0) br0: topology change detected, propagating br0: received tcn bpdu on port 1(eth0) br0: topology change detected, propagating br0: received tcn bpdu on port 1(eth0)
where packets are stored when they're received. After some googling, I think perhaps libpcap with mmap would be the solution. Does libpcap 1.0 and above support mmap?From my understanding, mmap would allow me to directly access the buffer without having to explicitly copy the packet to another buffer for me to do processing. I would also appreciate it if someone can let me know where I can find examples of such applications, as many examples of mmap I've found involves mapping a buffer for file operations, rather than integrating it with pcap functions like pcap_loop, pcap_next_ex etc.
sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'soft' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard'
PC1 runs radvd to provide router advertisements to the network and a DHCPv6 server for stateful addresses.Each interface is configured on a separate subnet. PC2 runs a DNS server on eth0. PC2:eth1 is used as an IPv6 client for testing purposes. The connections from PC1 to PC2 are just crossover cables.I've created virtual machines of both PCs and have created 4 virtual adapters on the host machine for each of the local-only interfaces.Now I have this:
The bold number of 6.4 is the % of sever memory this process is using. 6.4 % of 512 MB of memory is about 32 MB of memory, so it appears that this isn't being limited by php.ini. Am I correct? This leads to the next question: Is there some way to limit the amount of memory a single suphp process can use? (Basically, something like the setting in php.ini which limits suphp processes in the same way.)
I have an application where I am sending data via serial port from PC1 (Java App) and reading that data in PC2 (C++ App). The problem that I am facing is that my PC2 (C++ App) is not able to read complete data sent by PC1 i.e. from my PC1 I am sending 190 bytes but PC2 is able to read close to 140 bytes though I am trying to read in a loop.Below is code snippet of my C++ AppOpen the connection to serial port