I am looking to build a Ubuntu based web filter. What we would like it to do is block access to certain sites for our company. We have had several employees get caught spending hours on end on gambling sites so we would like to restrict access to websites on a per user basis.What I am looking for is a piece of software or suite of software that can filter websites based on a blacklist/whitelist or category based scenario. I need to be able to authenticate users. For example I would like it so that when the CEO logs in he can go to whatever website he wants, while most other staff members are blocked from accessing things in the blacklist or categories.
I remember from a recent trip to a hospital that they had all internet traffic re-routed to their landing page and that you had to agree to specific terms on that page before you could do anything else. Something like that might be useful as well.
I have a very simple set up.With Network Manager I can have my laptop act as a router (sharing all connections).I also have apt-cacher-ng as a debian package cacher.I would like to set up iptables to filter only the urls that are meant for a debian package cacher. For example:I could use a "forward all" rule:
Code: iptables -t nat -A PREROUTING -i eth0 -p tcp --dport 80 -j REDIRECT --to-port 3142 Except then I would get a bunch of error pages every time I tried to do normal navigating.My question (again) is: Can iptables handle forwarding only on a specific url? If so how?Or is there another solution? (prferably without full fledged software like squid)
last friday my schools isp enabled iwsva - interscan web security virtual appliance from trendmicro. all traffic is now filtered, and slowed down. videos and other videostreams are now useless. after this we have had serious problems using our netbooks with ubuntu, and upgrading workstations with ubuntu has become very problematic, or almost impossible. i have to run the update process 6-8 times before all packages are updated.
what can i do to document the problem in a good way. the isp tells me that there are no problems, and that it has to be a linux problem. (ever heard that one before?) the ltsp-clients on my debian servers does not have the same problems, maybe because they are using the proxy on this server?
We have MS Exchange email server with postfix/amavis on FreeBSD as proxy for anti-spam and anti-virus. We use our own developed CRM and trying to implement such functionality that when CRM manager send email to CRM client or vice versa through email client like Outlook (no plugins for CRM) this message would be shown in CRM. There is already a solution to parse emails from specific emailbox and show it in CRM.
What I need is to create postfix email filter to check if sender email is in file CRM_managers_emails and recipient email is in file CRM_clients_emails then forward that email to [URL] Filter should check both incoming and outgoing messages. I will auto generate CRM_managers_emails and CRM_clients_email files containing all such emails taken from CRM database. I know i should use something like Postfix After-Queue Content Filter but most of examples are for blocking emails.
I'm deploying new ubuntu server which should act as a router. I've already set up the NAT for local network, and also did some shaping for different groups of users, but now I'm facing new problem.I need to make a scheduled URL filter. I know it's not a problem with cron and simple script, but maybe there is existing way to do that? And also, I need to make statistics on web-traffic. I need to have list of URLs visited by users (source ip, destination url). Is it possible with iptables? or with any other software but without using proxy servers.
I have a system running at home that uses Getmail to retrieve mail from my ISP's pop server. Dovecot then offers that mail over IMAPs to my desktops running Thunderbird.
The reason I have resorted to using Getmail is because I don't have a static IP (from my ISP) for my server, and thus this server doesn't act as an MX.
I have implemented Spamassassin in my Getmail script as described here.
From further research, I understand that in order to fully utilize Spamassassin 's potential, I have to resort to training it with SA learn.
Currently I still receive spam messages, but 50% of spam is marked as ****SPAM****, and the other half is not marked at all.
My question is this:
1) How do I get getmail to move messages marked as spam by spamassasin to be moved to a JUNK folder within my mailbox automatically?
2) I thought of creating a folder where my users can move messages they deem to be spam, and set up a crontab script to invoke salearn regularly on this folder to get the bayes engine to learn from it. Is this the correct way of doing it?
I've been looking for a way to filter out emails on my postfix server by GeoIP data. I couldn't find anything that fitted the bill so wrote my own in Python as a postfix policy. Thought I'd post it here to see if it's of any use to anyone else. I've attached a tar of the files as the Python formatting will get mucked up by the forum code. Comments/improvements are welcome (be kind ) First file is : policyd-geoip which is owned by root:root and placed in /usr/bin with 755 perms
I am running a basic squid + privoxy combo for web caching/filtering proxy and it works fine. I'm basically running a stock config w/ a few minor edits to allow the relevant hosts access etc.now I am trying to find an easy way to specify privoxy to not filter a specific site (which it breaks). I have tried wading through the privoxy manual + google but I find the config file incredibly complicated. Are there any experts out there who can tell me: what is the easiest way to tell privoxy to 'pass through' a specific website?
I have setup and apache server on my pogo plug pink. I would like to setup a a web based file manager. I Googled it and tried a few of the top few but none of the worked that well.
I have installed 9.10 server on an old machine at home that I want to use purely for managing any software downloads.
I used to use the firefox add on DownThemAll on my ubuntu desktop environment so I am looking for a web based download manager for the server that has similar features (username/password restricted downloads, scheduling, pausing/restarting downloads) as DownThemAll.
I basically want to be able to add a bunch of downloads to a list and the server then downloads them. I need to be able to save a username/password combination for certain sites. I would also like to be able to see progress on the downloads and pause/resume them
We need to only allow access to a web page only to some ips. We want to do this via Apache but it seems like the standard way of denying all and allowing some ips doesn't work because we mount a webapp using tomcat.
I want to connect to the same machine that that I have OpenSSH server on which uses keys and I have disabled password-based logins (for ssh). Apparently, this also affects SFTP which makes sense. How do I setup SFTP to use key-based authentication?
What ubuntu server setup will work the best on a non web mostly data based via sockets (mysql php phpadmin) will be the easiest to use and what sockets to use? Lamp?
Know any documentation or software packages to do a open source "File Hosting" or also known as "one-click hosting" server.I want to create my own private secure site to easily have clients download sensitive files. If it could be setup to use SSL that would be great.
I need to setup an linux cluster ..so i prefer ubuntu because of support and i personally i use ubuntu.. and can any one explain in breif ..what all the things needed to setup an ubuntu based cluster my configuration for each node will be (totally 6 nodes) core2 duo with 4 gb ram i need 4 nodes and 2 for load balancing..
I am moving some websites from a shared host to a newly configured Xen based VPS with Ubuntu 10.04. A steep learning curve but I am getting there! I have a couple of sites installed on the new server and want to test them out, from what I understand I should be able to edit /etc/hosts on my desktop computer (running Ubuntu 10.04) which should automatically redirect my browser to the new server. But it doesn't seem to work.
This is what I appended to the /etc/hosts file: 12.34.56.78 www.xxxx.com
I thought that this should send any requests for www.xxxx.com to 12.34.56.78? Some googling indicated that the browsers I am using may be using DNS caching, ignoring the hosts file. I've tried clearing the cache with Chromium 7.0.517.44 (64615) and Firefox 3.6.12, and even installed a plugin in Firefox to block DNS caching.
how do I tell grub to stop trying to boot based on UUIDs? I've been dealing with this for about two years now. I have this one system, a PowerEdge 1550, that for some reason WILL NOT boot using a UUID, and will only boot if I edit the first entry in grub, remove everything to do with UUIDs (completely delete the search line, replace root=uuid=[uuid] with root=/dev/sda1). I can't seem to make these settings stick and now with 10.10 there's no longer a menu.lst to edit. So, in short, I need grub to stop trying to boot based on a UUID. Absolute paths, baby. That's what I need.
I have a script I want to execute after successful login for users who are members of a specific group. Ubuntu server 64 bit, 10.04. Where could I accomplish this?
i need to make an file server.but I'm not gonna ask how because that's a piece of pie.The thing is i need to make an ubuntu/xubuntu based file server.For windows users.But not just a regular file server. A File Server with accounts. That when you try to access the file server in explorer you need to provide a username and password.And that you( as the administrator) manage the privileges for the server/folder/file. so with different permissions.
i keep receiving this damn annoying spam: Quote: No Cost Website Analysis and Ranking Report Rank mysite.com at the top of the 3 major search engines Do you want to know why your competitors are showing up at the top of the three major search engines and how you can rank on top? To increase the number of visitors to your website it is crucial that you have a top search engine position. If you are like most people, you will spend countless hours trying to get traffic to your website.
Our search engine optimization experts will run a detailed website ranking and analysis report showing you exactly where mysite.com currently stands in all the major search engines along with the recommendations of how we can increase your ranking on Google, Yahoo and Bing. For Your No Cost Analysis Report and More Info Please Click Here
[Code]...
i get about 30 of those a day, everyday, for a long time. ive asked the ISP to stop it (theplanet.com) they are not responding and the spam continues. i was thinking is there a way that i can tell postfix to reject emails that contain "No Cost Website Analysis and Ranking Report" in their subject? there must be a way but i dont know how.
For the few thousands of people who have a little too much time on their hands who create personal web servers and host them using their local ISP on dynamic IP addressees.... what does the future hold for me.. {COUGH} ... I mean, them when IPv6 rolls out?
I have setup a home based web server to host a photo blog for myself and my friends. I will be running wordpress and possibly a phpbb3 forum. I'd like to open this to discuss server administration, server setup, and server maintenance. However, I have a pretty good start on all of those but serving a domain name to my static ip. Here my static ip is 24.10.202.144. I registered a domain through [URL]... It appears that I have the domain working to forward to my ip. However, I am still getting this output file from apache.
[code]...
I have tested the domain name across a few different computers on different ips. It works appropriately. I just want to make sure I have it set correctly on the (apache) server side of things. Then I can get more into Zone Editing etc.
Tomcat6 is built in server with it but i dont know how to use it .I am doing a dynamic web project based on servlet and eclipse galileo is being used as IDE for it.
I am working on a set up, and I like to be able to toggle compiz on and off, or toggle dualscreen on and off. The thing is, I also have conky on my desktop as well as a terminal window embedded in my desktop (that requires compiz). So, when I turn comiz off, or resize my desktop, I want to be able to reposition conky/embedded-terminal and the terminals position is relative to my conky position and the size of my virtual desktop.
I can do this all fine, except that to reposition the terminal I need to kill it then reopen it. But if I kill gnome-terminal it kills ALL gnome-terminals instead of just my embedded one. How can I specifically close my embedded one and leave any others untouched? Lets say that the title of my embedded terminal is "trans777"Also, the trans777 titled gnome-terminal will be killed when compiz is not running.
I am thinking about writing a script to search through specific folders on a Linux based server (Centos 5) and report any changes made to any file on that server.Like a detailed change log with the date it was changed and the difference in size. Dose anyone know if such a script exist?
I notice that on DistroWatch, the descriptions include "debian based", "arch based", etc. For the newbie, is there any practical difference between the different bases, or is it a Coke vs Pepsi type question?
Is a distro based on Debian say, easier for a newbie to learn or work with than say one that is FreeBSD based?
I've got some source code that loads and plays WAV files through the OSS drivers (and USB audio). The code works even under 2.6.34. Is there any benefit to rewriting the code using the ALSA system?