i have centos os squid 2.6 version,i have to configured squid to restrict some ip to 10 kb upload for that i set request_max_body_max_size but this directive is applicable to all ips but i want to limit uploading for some paricular ips.
I have Squid Cache: Version 2.6.STABLE21, installed on RHEL5. as well as configured SARG. It provides downloading detail of squid users. but does not provide uploading detail.I want to maintain uploading record of my users.(e.g which user upload which file/data)
Is there any way of limiting total download from a particular ip in a month with the help of squid proxy. Through delay pool I beleive we can limit the download speed.But I want total download limit for particular ip or user
I m using squid 2.7 Stable 9 and Dansguardian 2.10.1.1, i have compiled both squid and dansguardian, i have enabled follow_x_forwarded_for in squid to make clients IPs visible to squid, i have also set x_forwarded_for=on in dansguardian, this is working fine and clients ips are visible to squid. Now i want to set down-loadable file size limit upto 50 MB in squid by using the acl reply_body_max_size 52428800 allow mynetwork for every user except few users the above acl is not working properly. mynetwork is our private network which is 192.168.0.0/16.
When i set the acl reply_body_max_size 52428800 allow localhost . it works fine but only for localhost. I want to allow upto 50 MB down-loadable file size to every user in my network except a few users whom will have access upto 500 MB down-loadable file size.
I want to limit bandwidth for downloading files with squid. I want to reserve bandwidth for other traffic (esp. web browsing). I know about delay pools but I don't understand well. Some users use download managers to download large movie files. I don't want to block downloading but I want to give them limited bandwidth. May be 5KBps or 6KBps because I have only 512Kbps (64KBps) connection.
I running Squid 2.7. I using NCSA veryfication so only registered users may use proxy server. I want to allow only 1 connection to proxy for 1 user. Why not per ip ? Alot of users have dynamic IP's and i want to avoid password sharing beetween users. Any idea how to do it in squid ? Mb other autentyfication system connected to squid granting access to proxy server ?
How to set download limit using SQUID? I want to specify the download limit for a particular list in MB. Is it possible to limit bandwith for some group of machines in network?
I just want to place upload/download limit to each ip address ( say 20MB per day for each ip or specific ip) using squid proxy i tried through delay pool but it control only download/upload speed not place any download/upload quota restriction. following are my codes I have 1Mbps line
delay_pools 2 delay_class 1 2 # here 700kbps(87Kbytes) Net-Total usage limit, with 50Kbytes per user
[code]...
I also used
"reply_body_max_size" parameter to control upload limit.
But
"request_body_max_size" only restrict download based on size.
how to setup quota ( download quota per day per ip)
How to set download limit using SQUID? I want to specify the download limit for a particular list in MB. Is it possible to limit bandwith for some group of machines in network?
I am using squid to controlling access to the internet all is working fine expect one of the user who is using outside organization portal to connect internet. But whenever he tries to enter in the portal by typing (EXAMPLE)url. Permission denied error from squid occur.
How can i allow this portal in squid. So squid will allow this to access.
Let's say we have machine 1 and machine 2. Machine 2, which is linux based, is receiving a file (let's say cat.jpg sent via SCP from machine 1 (doesn't really matter which operating system..) How can I know in machine 2, that the file has finished uploading from machine 1 ? Please note that by "knowing" I mean by running a command (via a cron that runs every minute), and not by an event driven way ("make the scp issue a command when it is finished uploading"). Another limitation I have is that I can't change anything in machine 1 or in the way it SCP files.
If it really matters: the files being transferred are video file, MP4 extension. machine 2 is ubuntu 10.10
authors recommended to upload pages of a site in a condensed form to the server to save in uploads and downloads by visitors. This involved removing all unwanted spaces and making one single string whenever possible.Is this approach still encouraged?Is there software that does that and respects the line-oriented scripts like Javascript embeded in HTML code?
I am running Debian 5.0.5 and want it to convert a video file after upload, then update in MySQL if possible? I have tryed ff-mpeg and yes I could convert to FLV but the Pixelation where very bad and and if I closed the browser it finished the conversion but didn't updated it in the database.
I now have a .tar.gz that has been created during my script, and finnaly i need to upload it via ftp to my computer, but i am unable to because i keep getting the "Connection refused" error, despite the fact that all the syntax appears to be right. in a bout of frustration, i have deleted the code and now no longer have what i had written.. so could someone please let me know the syntax to upload a file to a server? I am aware that other threads have been started about this issue but I am trying to ftp from an iPhone to a computer...
I got an error while uploading a file named "xxxxxxxx.php"to linux server
Error in detail Upload of file 'xxxxxxxx.php' was successful, but error occurred while setting the permissions and/or timestamp. If the problem persists, turn on 'Ignore permission errors' option. Permission denied. Error code: 3 Error message from server: Permission denied Request code: 9
I had configured openvpn (2.0.9) on my Centos 5.4 machine.I want to route all client traffic through my gateway.So i had enabled (push "redirect-gateway ") in server configuration.Now the problem is the upload speed is very low while connecting through vpn.Normally i got 700 kbps download and 650 kbps upload speed.But while connecting through Vpn the upload speed is only 110-130 kbps but download is almost stable.Here is my Server and client conf Files.
sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'soft' Dec 28 22:42:29 yn54 sudo: pam_limits(sudo:session): wrong limit value 'unlimited' for limit type 'hard'
I've got several large files sitting in my linux hosted account that I need to upload to my S3 account. I dont want to download them first and then upload it into S3. Is there any way I can "upload" it via the linux command line environment? Or access it via a website working with lynx?
my webpage PHP below. I would like to enter "Hello, in the main inputbox field" below (You are editing: textfile.txt) and click "SAVE" directly from command line.
Sort of : wput_php "hello, in the main inputbox field" click save, and here is it. the text would be uploaded.
I am looking for a file sharing program to install on my dedicated server that will allow me to upload large MP3 files and allow my clients to download them. these files are recordings of counseling sessions for families who are seeking help for their children.
What I am looking for is similar to the system this company uses [URL].
I'm having difficulties with uploading files to a CentOS-server with vsftpd. I have the exact same configuration on a Fedora10 and there I have no problems...
I have a VPS server with 512 MB memory. The php.ini is set so script memory limit = 16 MB. However, I have noticed in my top report, instances like the following:
The bold number of 6.4 is the % of sever memory this process is using. 6.4 % of 512 MB of memory is about 32 MB of memory, so it appears that this isn't being limited by php.ini. Am I correct? This leads to the next question: Is there some way to limit the amount of memory a single suphp process can use? (Basically, something like the setting in php.ini which limits suphp processes in the same way.)
My squid server works fine in fedora 11 system . Is there any web like interface for admins to create,change,modify users of squid and to view their logs.
I would like to ask some help and tutorial for setting up and how to configure squid proxy server in my (Home PC Server). I am a newbie in Linux Centos. I already installed in my system the CentOS 5.5 . Now, I want to configure it as my internet server, all of my 4 system running in Windows including the laptop I want to connect through my CentOS pc with username authentication. I assign all IP address by static. see tthe attachement in my set up. [url] I just want to know what I need to change and add in my squid config file. And how can I configure properly my CentOS with 2 LAN card as internet server.
I would like to limit any process from using more than 500 MB of RAM. AFAIK this is done using RSS in /etc/security/limits.conf but the process called gnome-panel apparently is using 618436 kB of VmRSS. How can this be ?
I have a problem with both genisoimage and mkisofs. Both of them are limited to 8 characters. There are very many options for them. Which one would remedy the issue?