Security :: Forbidden Access To Some Websites It Says This Domain Is Blocked?
Sep 17, 2010
I don't know is this the right place to ask, but i must ask some questions Here's my problem.I'm a student in highscool,and here we use Linux(ubuntu) OS .Every classroom has like 30 PC's connected with the main computer(the teacher's one) so....3 days ago we were forbidden access to some websites it says This domain is Blocked.By the way the Linux version installed is 7.04(feisty Fawn) i tried disable cookies that did not worked,also tried to whitelist some website,that also didn't worked out
I'm using Ubuntu x64 (dunno which version, but I don't think it matters) and I'm concerned about security with PHP.I remember using lighttpd and I had some mystic configuration and the secuirty was perfect for me - if one website gets hacked then the others are still safe (kinda).Now with apache2 if I enable safemode I'm still able to go outside web directory and actually I can go really far untill user/group matches.I tested the system with r57shell and I was able to mess up other websites.Is there a way to disallow access to other websites?
I'd like to be able to limit access to a particular website, based on the time of day. I would also like to be able to password protect this if possible.So for instance, from 7am until 10pm daily, I can access URL... but after 10pm it redirects to 127.0.0.1 or something. And this configuration be protected by only allowing a certain user (other than root) to change the config?
I have been trying to get Squid to work so that I can restrict access to a particular web site during certain hours every night. I can't seem to get it working, however. I am still able to access the site. The following are the relevant lines from my squid.conf file:
acl restricted-domain dstdomain "/etc/squid/denied_domains.acl" acl test time 19:00-20:00 acl bedtime time 22:00-23:59
Have no idea what I am doing operating a server. Our programmer got a new job and I am the one who has to take it over. Everything was fine til yesterday. You see I made a control panel to easier update the site www(dot)discoverysound(dot)com
but yesterday when I went to update the site I got an error called fopen. I thought I fixed it (and boy did I ever) but now I cannot get to my site because it says Forbidden You don't have permission to access / on this server. Apache/2.0.47 (Unix) mod_ssl/2.0.47 OpenSSL/0.9.7a DAV/2 PHP/4.3.3RC4-dev Server at www(dot)discoverysound(dot)comPort 80.
So I'm looking to host potentialy three different...sites...I want to call them from the same machine. One is a radio station and the other 2 are just straight websites. So I wanted to know if I had to use different machines to accomplish this or can I alter maybe the http conf file or what. I did find this in the forums but wasn't quite sure if this applied to me. [URL="http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/"http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/[/URL]
I have an old FC2 box running Squid version 2.5. It has been running since 2003 so I am in the process of replacing it. I have a new machine with FC11, iptables, and Squid 3.0 installed.
On the old machine I use iptables to intercept Port 80 traffic and send it to Squid. By default I block all internet access and allow only sites that are in an Allowed_Sites.txt file. Within Squid I also have statements to allow certain users to bypass Squid based on their IP address.
I have set up the same thing on the new box. I have iptables intercepting the Port 80 traffic and sending it to Squid. That is working because if I remove the redirect statement from iptables all internet access is blocked.
The problem I am having is that Squid is not blocking any websites. It acts like the ACL is set to http_access allow all. I have worked on this for several hours and am stumped.
I'm having some problems allowing websites with the word essex in the domain name. I've been running dansguardan for some time now and have managed to make rules to allow and disallow sites, but I've now hit a brick wall on the latest request from staff here. which config file i need to edit.I've tried adding *essex*.* to the exceptionsitelist, but these sites are still blocked. Example sites are [URL]
before i using fedora 14, i was used fedora 8.... no problem in setting web server, all client (windows XP) can access my resources in var/www/html/index.html but after upgrade to fedora 14.... this is a problem
"forbidden you don't have permission to access.... apache/2.2.16 (Fedora) server at 192.168.1.1 port 80" i had disable firewall also set SELinux to permissive
Nagios - forbidden you dont have access to this server I had made following changes on httpd.conf file: ScriptAlias /nagios/cgi-bin/ "/usr/local/nagios/sbin/" <Directory "/usr/local/nagios/sbin/"> AllowOverride AuthCongig Options ExecCGI Order allow,deny Allow from all </Directory> Alias /nagios/ "/usr/local/nagios/share/" <Directory "/usr/local/nagios/share/"> Options None AllowOverride AuthCongig Order allow,deny Allow from all </Directory>
I am trying to browse in /var/www/directory and it giving me a following error. Forbidden You don't have permission to access /directory/ on this server. Apache/2.2.14 (Ubuntu)Server at localhost Port 80 I have just copied that directory to /var/www. What might be wrong.
The company I work for, as usual, is Microsoft-centric. I'm attempting to integrate my Ubuntu server into the domain to allow domain users to authenticate to the server and access file shares using Samba. Here's my current configuration:
I am trying to configure CUPS and when I click on Adminitrator to Add A Printer, I get Forbidden after entering my root username and password. Anything I'm doing wrong?? In the past, it was just a matter of putting my username and password and walking though the installation. The printer is connected to the computer via USB and is being seen on the machine.
I don't have permission to access [URL]. I can access [URL] only local. I configured local and network access in /etc/httpd/conf.d/nagios.conf. My access_log: "GET /nagios HTTP/1.1" 403 334 "-" "Mozilla/5.0 (X11;U;Linux i686; en-US; en -US; rv:1.9.1.9) Gecko/2010100330 Fedora/3.5.9-1 fc11 Firefox/3.5.9" "GET /vavicon.ico HTTP/1.1" 404 335 "-" "Mozilla/5.0 X11;U;Linux i686; en-US; en -US; rv:1.9.1.9) Gecko/2010100330 Fedora/3.5.9-1 fc11 Firefox/3.5.9" My error_log: suEXEC mechanism enabled (wrapper: /usr/sbin/suexec)
Digest: genarating secret for digest authentication done mod_python: Creating 4 session mutexes based o 150 max processes and 0 max threads mod_python: using mutex_directory /tmp Apache/2.2.15 (Unix) DAV/2 PHP/5.2.13 mod_python/3.3.1 Python/2.6 mod_ssl/2.2.15 OpenSSL/0.9.8n-fips mod_perl/2.0.4 Perl/v5.10.0 configured -- resuming normal operations Permission denied: access to /nagios denied File does not exist: /var/www/html/favicon.ico Directory index forbidden by Option directive: /var/www/html
installation moodle on Ubuntu 10.04.1 LTS.at localhost show message It works!at localhost/moodle show message Forbidden You don't have permission to access /moodle on this server.
I have installed Redmine according to the instructions on [url] according to the mod_cgi instructions which seem to be the easiest to follow.
The script installed in the installation directory was called 'dispatch.fcgi', and not parsing when I visited the page through the browser, just displaying the contents of the script. OK, so I installed the apache mod libapache2-mod-fastcgi using apt-get. Then 'a2enmod fcgid' followed by an apache restart.
Now when I visit the Redmine folder through the web browser I get 'Forbidden You don't have permission to access /redmine/dispatch.fcgi on this server.'
I checked to ensure www-data owns all files and has correct permissions to the folder. But this doesn't seem to work.
Using httpd - everything usually works but I have written a cgi script and placed in cgi-bin but i cannot access this dir and keep getting forbidden messages. I have tried adding my root and usual user to the /etc/group file against the user stipulated in the httpd.conf file but i still can't access the cgi script i wrote (in browser - localhost/cgibin/myscrippt.cgi) when i place it in the cgi-bin under htdocs. i have also chmod 755 directory and files.
I've setup the Uncomplicated Firewall (UFW) on Ubuntu 10.04 LTS and blocked an IP address. UFW status shows that the firewall is active and the IP in question is denied. The issue is that I'm seeing the blocked IP address in my Apache logs.
If I use my current iptables firewall settings my php includes from my own site get blocked by the firewall. If I use accept all tcp and udp it still does not work. If I use a default policy of accept all, then the php includes work which is not safe. I am running a drupal 6 site on a centos 5.3 vps with apache 2.2.0 and php 5.2.6 w/ safe mode on. He is my current iptable configuration.
We have a spam in our network and we installed antivirus in all our systems and cleaned the virus from all pc's after that i had removed my ip from the database of blocked ip's but still my ip is blocked for sending spam i don't know from which pc the spam is going on the internet.
so i have a question that my proxy server is redhat linux and as a newbie i don't know the command's to find out which pc is creating large bandwidth to the internet. If you tell the command how to see which pc is sending spam then i will discard that pc. Also i want a strong firewall to stop spam activities.
I can't seem to remotely SSH or VNC into my machine. If I'm on the LAN and try accessing via LAN IP, it works fine. If i go in through a remote address (my dyndns) or even my home IP, I can't connect (yes, all of the ports are forwarded, I've triple checked this multiple times). Interestingly enough, port 80 works just fine. It would seem as though some sort of firewall is blocking me. I've done this plenty of times before with various machines, and this has me quite perplexed.
Basicly just installed a fresh version of Cent OS 5.4 with apache httpd installed automaticly during the installation. The http daemon is running and when navigating to localhost i get the welcome cent os apache page thingy. The problem is when i put an index.html file in /var/http/www then try and navigation to localhost I get a 403 forbidden error.
I am on mac os x and whenever I make aliases in my apache config file to point to folders outside of main web document root folder, I get access forbidden. I have chmod -R 777 the whole destination directory (to test it) and still no go. Access forbidden. I have also tried pointing directly to [URL].
In the office there is a local network with samba+openldap PDC. The local domain name is company.net. The company desided to create a corporate Website on a remote hosting and desided that the site's domain should be company.net which is same as local network's domain name. So now it is not possible to reach that corporate website from within the company's local network because, as I guess, bind9 which is installed on above menioned PDC looks for company.net on a local webserver. Is there a possibility to let people from this local network browse the remote site?
I have installed MoBlock as instructed here: [URL]
After installation I created my own list file in /etc/blockcontrol/custom-blocklist.p2p and have the following uncommented at the bottom of /etc/blockcontrol/blocklists.list:
Code: Yahoo:98.137.149.56 Google:74.125.47.147 When I do:
[Code].....
Recently I just noticed that the locallist rules seem to have no effect. I will always get "destination port unreachable" even if the locallist entry in blocklists.list is commented out.
However, whenever I try to browse to that IP, even when blockcontrol is on, even by typing the IP into Konqueror (not the domain name), it lets me go there every time. How can I know that my other applications will not to do the same thing? How can I lock this down and test it empirically to be sure?
On April 10, 2010, I upgraded some packages on my Ubuntu 9.04 server. This included an upgrade to "ufw 0.27-0ubuntu2". I rebooted the server, and all appeared to be fine.
Now I've noticed that UFW is not logging blocked packets since that reboot. It used to do this. It is still logging the allowed packets that I've configured it to log.
I come to Ubuntu with the notion that it is much more secure than Windows. In XP I had an anti-virus, third-party firewall and sundry softwares against spybots, rootkits etc. The anitivirus blocked the suspicious web pages while browsing. I generally avoided public networks, carrying a portable internet device Do I need similar stuff with Ubuntu.