I was wondering how to block attachments to gmail. I am running squid 2.7 stab9 with dansguardian 2.10, users authenticated from LDAP. I have configured the POST restrictions in Dansguardian which does block all attachments to hotmail/yahoo etc etc but attachments to gmail continue to upload.
I manage a linux-based network, where some projects are currently under development. Our IT policy states that any email attachment shall be encrypted using GPG. Can I block other attachments using a firewall?
Note: Currently our mail server is not in campus. So I can only use a firewall for this security issue.
how to disable the gmail chat? My means to say that when we login to gmail , after that the chat will open, I want to disable that chat. am using Redhat 9 and squid stable 2.5 version. I have tried the things mentioned below, but chat is still working.
My Problem is: I want to stop gmail access without blocking https. Yes in my squid proxy normal http://gmail.com is not accessible. But gmail recently started https service by which user can still get access to gmail. I DONT WANT TO STOP https CAUSE ITS BEING USING BY MY COMPANY GOOGLE MAIL PROGRAM.
I have blocked both gtalk and gmail chatting for total network using both squid and iptables. I want to give access for gtalk and gmail chatting to some user. How I can do it.
Chrome crashes when I try to send attachments using Gmail, instead firefox works fine. Several users have the same problem, but I could't find a solution.I use openSuse 11.3, Chrome 7.0.517.44 and the lastest flash version.
My Problem is: I want to stop gmail access without blocking https. Yes in my squid proxy normal [URL].. is not accessible. But gmail recently started https service by which user can still get access to gmail. I DONT WANT TO STOP https CAUSE ITS BEING USING BY OTHER PROGRAMS.
The problem is here:When I was open gmail in my system Certificate Error is coming. The error details:
This Connection is Untrusted You have asked Firefox to connect securely to url, but we can't confirm that your connection is secure. Normally, when you try to connect securely, sites will present trusted identification to prove that you are going to the right place. However, this site's identity can't be verified.
What Should I Do? If you usually connect to this site without problems, this error could mean that someone is trying to impersonate the site, and you shouldn't continue. code....
This used to work in Fedora < 15. Whenever I want to send an e-mail using Gmail's SMTP server for the first time in a given session, the KWallet system asks if I would not rather migrate the password and store it encrypted inside a wallet. Every time I concur and every time an error message pops-up, as soon as I try and send another e-mail. The attached snapshot shows the error message I get. It doesn't matter how many times I select "Store Password". Every time the ordeal is the same. Up to Fedora 14 this method worked all right, but now it fails.
I recently received an email from a friend without subject and just a link. Since we do that a lot, I clicked on it. I was taking to a website that looks like a phishing site and my computer hard drive started working feverishly. I closed it quickly.
First, I want you guys to be aware of this thing since it seems to be fairly new. Second, I want to know if I have been compromised. I already changed the password on my gmail account and I accessed the site using Ubuntu and Firefox.
second time clamav detects the malware on laptop underubuntu:winnow.compromised.ts.jsexploit.5.UNOFFICIALwinnow.spam.ts.domains.158.UNOFFICIALgspace.js: winnow.malware.cm.miscspam.387929.UNOFFICIALwhat does this mean, is it serious and what is the origin of this infection?
on my linux server i have many websites but with difrent ips address, is some way to i can block all the ips with many connection (100+) just from my website not from all websites
What is the easiest way to block one specifiek web page?Can I use the file /etc/deny host, or should I use another program to do this?I have already search the web andfound iptables, but that is to difficult for me, and I found squid
I'm trying to block an incoming URL. My ISP is hijacking 404 pages and annoyingly changing the URL line in the browser and flashing all sorts of popup ads. I just need it for incoming URLs which my router doesn't seem to handle. I'd prefer something packaged with Ubuntu 8.04, but anything simple will do. I know in KDE I could edit the kdeglobals file with:
Im running Centos 5.4 with a sftp server, and I�d like to allow all 172.16.0.x ip and 192.168.0.x ip and block everything else. Does someone have a good way to do this with IPTables or any other opensource FW?
What's the best way in centos to block a user from accessing mysql. I don't want him to be able to run the mysql command, so just putting passwords up in mysql is not good enough. Mysql is running ad user=mysql, and i added the user in a different group by he is able to access mysql by typing in the command.
How can i block this command being availible for this user.
I am looking for an answer about how to allow just one trusted DHCP server and block others ?I am using Centos 5.5, iptables and dhclient.I have read that it is impossible to block DHCP Replay using iptables: URL...So how can I do that ? Maybe another dhcp client?
here i would like to inform you that my company ask me to block facebook site at the work time it is: 08:00-12:30 and 14:00-17:30 in proxy server.i'm still confuse how to do!i need the details configuration in squid regarding how to block the sites
Is it possible to block a subdomain or a one lower level directory URL access from other hosts or network ? I have a site running on my server and i want to block the particular directory under the domain, with the exception of loopback access? I mean the directory must be accessible from loopback/localhost.
[url] on port 10016(expect loopback) [url] on port 10016 (expect loopback)
use any browser and go to gmail. i can't get gmail to load. if i browse on videos, links to videos don't work. they appear to be taking a long time to buffer, then a message appears saying to try again. sometimes, links on google search pages are all dead. yahoo mail loads as far as logging me in, then i get a page with all dead links.
i'm using a laptop at several locations that have free public wifi. at some places, everything works perfectly normally. at other places, google and yahoo are almost useless. the "bad" places work just fine under ms windows.
so i figure it's a networking/handshaking issue, but that's all i can figure. i need help to troubleshoot it further. if it's happening to me, it has to be happening to other people, and i consider it a major hassle.
My college internet providers shifted to a different setting.. defined in the changing lan proxy settings on the user end. I.e.ip - 192.168.0.200 and port to 3000. This to reduce misuse of the net connection ..The thing is, since then, every other website is working, except GMail. It isnt blocked, the Nebero page shows up for blocked sites. When attempting to open gmail, the default internet page shows up for unavailability of page owing to no or slow response from the server, or the firewall or proxy settings not appropriately configured. I cannot open gmail, the connection on empathy im. I cant ping gmail either. No response. What can be the problem.
I recently set up a web server at home, using a non-standard port, due to my ISP blocking 80. I just checked my log files, and I see a TON of entries indicating that a file was not found "proxy-1.php", "proxyheader.php", etc. I do not have these files, not intend to have them as part of my website. I did a whois looking by IP address for several of these, and they all seem to come from an ISP in China. Is there a way to BLOCK any IP address outside the US (that is somewhat simple to do?)
My question is how to block a subdomain of a site. To make it as clear as possible, I'll give an example. I am regularly entering this arbitrary site [URL] which redirects me to this page [URL] and this index.html takes an image from a subdomain which is a subfolder of itself, that is: [URL]. What I am asking is blocking the images to be taken, but not the main page itself, i.e. to block www.somesite.abc/images/ without blocking the overall www.somesite.abc.
My idea was to use the /etc/hosts file by redirecting to loopback address: Code: 127.0.0.1 www.somesite.abc/images But it looks as if it doesn't affect things at all. Should I use it another way? Modifying /etc/hosts.deny maybe useful?