Im using squid3.0 stable25 in fedora11 RAM512MB, and have 25 users. Problm is it uses full ram after about 30 minutes from booting.. when i upgraded it to 1GB the whole 1GB is using after some time... i think its due to caching of squid, how to disable it?
I am using squid to controlling access to the internet all is working fine expect one of the user who is using outside organization portal to connect internet. But whenever he tries to enter in the portal by typing (EXAMPLE)url. Permission denied error from squid occur.
How can i allow this portal in squid. So squid will allow this to access.
My squid server works fine in fedora 11 system . Is there any web like interface for admins to create,change,modify users of squid and to view their logs.
I would like to ask some help and tutorial for setting up and how to configure squid proxy server in my (Home PC Server). I am a newbie in Linux Centos. I already installed in my system the CentOS 5.5 . Now, I want to configure it as my internet server, all of my 4 system running in Windows including the laptop I want to connect through my CentOS pc with username authentication. I assign all IP address by static. see tthe attachement in my set up. [url] I just want to know what I need to change and add in my squid config file. And how can I configure properly my CentOS with 2 LAN card as internet server.
I'm using squid for proxy server in FC6. I'm also using squidGuard for web-site access restriction. I want to do some exception now for website access. For example, squid user1 with ip 192.168.7.10/32 shoud not access facebook.com while all other squid users with ip 192.168.7.11/32, 192.168.7.9/32 and so on... can access facebook.com since facebook.com is not listed in squidGuard .db files
but it give me error as like: - (This is the output of # squid -k parse) aclParseAclLine: Invalid ACL type 'arp' FATAL: Bungled squid.conf line 1234: acl block arp 00:13:45:d3: 24:e4 squid Cache (Version 2.5.STABLE6) : Terminated abnormally
i have installed squid 2.6 on my centos. i have writen a shell script to ping a network and write to a file. write '1' if network up and '0' if network down. After that,a perl script will read the file and do the redirection.Perl will redirect to a fix URL [URL] if the network down and do nothing when it up. i have put my perl script in squid.conf at url_rewrite_program /my_file_path.
below is my shell script for pinging:
Quote:
#!/bin/bash while [ 1 ] do HOST=143.148.137.134
[code]....
My problem is client browser are not redirect to www.google.com even the network is down. It should go to the fix URL when the user click any URL in network down situation.it just appear cannot resolve host.
Can't seem to access my squid server on port 80. I have port 80 allowed in the conf for this IP. apache is listening to port 80 but only on the 2nd IP. iptables is allowing through port 80 incoming nmap shows no ports open on 80 though:
Code:
Starting Nmap 5.00 ( http://nmap.org ) at 2009-08-20 11:19 BST NSE: Loaded 0 scripts for scanning. Initiating SYN Stealth Scan at 11:19
I have installed squid and dansguardian on my server, I also setup my iptables to forward port 80 communication to port 3128 (squid). I also have remove the comment on /etc/dansguardian/dansguardianf1.conf (line "bannedextensionlist") hoping that my server would block download. But it isn't, it still download file no matter I add in /etc/dansguardian/lists/bannedextensionlist. Oh yeah, I also add this line to my /etc/squid/squid.conf
I recently finished installing Squid 3.1.9 and I think I've done installing correctly with its feature of minor configuration changes. It accepted requests on port 3128 and created and the created some numerical or binary (I guess) files in /usr/local/squid/var/logs, My problem is how can I fully verify if the cache is really storing Internet Files? I've read some forums in the Internet replied to me to try the command: cat /usr/local/squid/var/logs/cache.logSo I did tried it and it gives me this output:
2010/11/11 18:04:49| store_swap_size = 0 2010/11/11 18:04:50| storeLateRelease: released 0 objects 2010/11/11 18:05:41| Squid is already running! Process ID 5458
i have centos os squid 2.6 version,i have to configured squid to restrict some ip to 10 kb upload for that i set request_max_body_max_size but this directive is applicable to all ips but i want to limit uploading for some paricular ips.
I am using Red Hat Linux and i want to to configure SQUID server on it. The server has two Lan cards.WAN eth0 192.168.1.200LAN eth1 192.168.1.210 I have a static IP from my ISP. In my private network i have 40 computers running windows xp and fedora ,ubuntu as a destop pc's from range of 192.168.1.10 to 192.168.1.40 In my server computer I can brows internet because it is directly connected with the ISP.For configuring squid should I first make this computer as a gatewat by packet forwarding i.e # echo 1 > /proc/sys/net/ipv4/ip_forwardand then configure for squid server
I configured squid with authentication required. I have 10 client using WinXP operating system. I want these clients can share their data, DVDwriter and printer.
i have squid 2.6 installed on centos i created acl which should apply to this ip range i.e 192.168.1.10-192.168.1.15 and excluded 192.168.1.13 how can i do
I have to use squid proxy server on fedora14 to develop the application/ software based on the Proxy sever analysing and count number of user and downloaded file size... .My main problems are, I'm not able to configure the access.log file in "Common Logfile Format" and how to develop the application
In Iran there is a famous "access denied page" that redirects you to a strange page with a lot of HTML errors and lol, telling you RTFM about ridiculous Internet laws.I want to filter the contents of the page, because the page IP, URL, ... are all unknown.I don't know much about squid configuration scripts.I can read but cannot write
I have been requested to deliver the logs for last friday, but the problem is that I have configured squid logs to be weekly, not daily. How can I extract the logs generated friday from the week file?
I have one Squid server which was configured to share internet on our networks192.168.20.0) PCs Serrver:192.168.20.25:8080It is working smoothly.Now my problem is that I has to configure one sub server that should allow internet to our 192.168.22.0 network with 3128 port. The sub server will get the connection from main server and it share the inter net to the internal network.
I am facing problem to access my network PC's and even ping. My network scenario is as follows. I am using squid 2.6 stable 21 on RHEL5. all other PC's on network (OS is Windows XP Professional SP2) are connected to internet through squid, authentication is ON on squid. All PC's on network (Win XP Systems)are assigned IP statically and Default Gateway is set which is Squid's IP. I want to access these PC's (Win XP Systems) mean share data between them. The problem is that i am unable to access and even ping these PC's.
I'm using squid as a proxy in our organization. My problem is whenever the squid cache folder size exceeds 3 GB our proxy server starts to hang. Is there a way to write a script that it will automatically delete the squid cash when it exceeds the 3GB limit.