Security :: Forbidden Access To Some Websites It Says This Domain Is Blocked?
Sep 17, 2010
I don't know is this the right place to ask, but i must ask some questions Here's my problem.I'm a student in highscool,and here we use Linux(ubuntu) OS .Every classroom has like 30 PC's connected with the main computer(the teacher's one) so....3 days ago we were forbidden access to some websites it says This domain is Blocked.By the way the Linux version installed is 7.04(feisty Fawn) i tried disable cookies that did not worked,also tried to whitelist some website,that also didn't worked out
View 7 Replies
ADVERTISEMENT
May 4, 2010
I'm using Ubuntu x64 (dunno which version, but I don't think it matters) and I'm concerned about security with PHP.I remember using lighttpd and I had some mystic configuration and the secuirty was perfect for me - if one website gets hacked then the others are still safe (kinda).Now with apache2 if I enable safemode I'm still able to go outside web directory and actually I can go really far untill user/group matches.I tested the system with r57shell and I was able to mess up other websites.Is there a way to disallow access to other websites?
View 5 Replies
View Related
Jun 11, 2010
I'd like to be able to limit access to a particular website, based on the time of day. I would also like to be able to password protect this if possible.So for instance, from 7am until 10pm daily, I can access URL... but after 10pm it redirects to 127.0.0.1 or something. And this configuration be protected by only allowing a certain user (other than root) to change the config?
View 7 Replies
View Related
Jan 21, 2011
I have been trying to get Squid to work so that I can restrict access to a particular web site during certain hours every night. I can't seem to get it working, however. I am still able to access the site. The following are the relevant lines from my squid.conf file:
acl restricted-domain dstdomain "/etc/squid/denied_domains.acl"
acl test time 19:00-20:00
acl bedtime time 22:00-23:59
[code]...
View 2 Replies
View Related
Apr 12, 2010
I want to block all websites except desired websites.
View 3 Replies
View Related
Feb 9, 2010
Have no idea what I am doing operating a server. Our programmer got a new job and I am the one who has to take it over. Everything was fine til yesterday. You see I made a control panel to easier update the site www(dot)discoverysound(dot)com
but yesterday when I went to update the site I got an error called fopen. I thought I fixed it (and boy did I ever) but now I cannot get to my site because it says Forbidden You don't have permission to access / on this server. Apache/2.0.47 (Unix) mod_ssl/2.0.47 OpenSSL/0.9.7a DAV/2 PHP/4.3.3RC4-dev Server at www(dot)discoverysound(dot)comPort 80.
View 2 Replies
View Related
Dec 13, 2010
So I'm looking to host potentialy three different...sites...I want to call them from the same machine. One is a radio station and the other 2 are just straight websites. So I wanted to know if I had to use different machines to accomplish this or can I alter maybe the http conf file or what. I did find this in the forums but wasn't quite sure if this applied to me. [URL="http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/"http://www.linuxquestions.org/questions/linux-software-2/serving-multiple-domain-websites-on-one-server-329914/[/URL]
View 1 Replies
View Related
Jun 11, 2010
I have an old FC2 box running Squid version 2.5. It has been running since 2003 so I am in the process of replacing it. I have a new machine with FC11, iptables, and Squid 3.0 installed.
On the old machine I use iptables to intercept Port 80 traffic and send it to Squid. By default I block all internet access and allow only sites that are in an Allowed_Sites.txt file. Within Squid I also have statements to allow certain users to bypass Squid based on their IP address.
I have set up the same thing on the new box. I have iptables intercepting the Port 80 traffic and sending it to Squid. That is working because if I remove the redirect statement from iptables all internet access is blocked.
The problem I am having is that Squid is not blocking any websites. It acts like the ACL is set to http_access allow all. I have worked on this for several hours and am stumped.
These are my Squid rules:
acl allowed_sites url_regex "/etc/squid/Allowed_Sites.txt"
acl manager proto cache_object
acl localhost src 127.0.0.1/32
acl to_localhost dst 127.0.0.0/8 0.0.0.0/32
acl SSL_ports port 443
acl CONNECT method CONNECT
http_access allow manager localhost
http_access deny manager
http_access allow Bypass_Users
http_access deny !Safe_ports
http_access deny CONNECT !SSL_ports
acl our_networks src 192.168.1.0/24
http_access allow allowed_sites
http_access allow our_networks
http_access deny all
icp_access deny all
htcp_access deny all
http_port 192.168.1.254:3128 transparent
hierarchy_stoplist cgi-bin ?
access_log /var/log/squid/access.log squid
refresh_pattern ^ftp: 1440 20% 10080
refresh_pattern ^gopher: 1440 0% 1440
refresh_pattern -i (/cgi-bin/|?) 0 0% 0
refresh_pattern . 0 20% 4320
visible_hostname FC11.proxybox
icp_port 3130
coredump_dir /var/spool/squid
View 2 Replies
View Related
Jun 30, 2010
I'm having some problems allowing websites with the word essex in the domain name. I've been running dansguardan for some time now and have managed to make rules to allow and disallow sites, but I've now hit a brick wall on the latest request from staff here. which config file i need to edit.I've tried adding *essex*.* to the exceptionsitelist, but these sites are still blocked. Example sites are [URL]
View 1 Replies
View Related
Apr 9, 2011
before i using fedora 14, i was used fedora 8.... no problem in setting web server, all client (windows XP) can access my resources in var/www/html/index.html but after upgrade to fedora 14.... this is a problem
"forbidden you don't have permission to access.... apache/2.2.16 (Fedora) server at 192.168.1.1 port 80" i had disable firewall also set SELinux to permissive
View 14 Replies
View Related
Jan 12, 2010
Nagios - forbidden you dont have access to this server
I had made following changes on httpd.conf file:
ScriptAlias /nagios/cgi-bin/ "/usr/local/nagios/sbin/"
<Directory "/usr/local/nagios/sbin/">
AllowOverride AuthCongig
Options ExecCGI
Order allow,deny
Allow from all
</Directory>
Alias /nagios/ "/usr/local/nagios/share/"
<Directory "/usr/local/nagios/share/">
Options None
AllowOverride AuthCongig
Order allow,deny
Allow from all
</Directory>
View 1 Replies
View Related
Jul 22, 2011
I am trying to browse in /var/www/directory and it giving me a following error. Forbidden You don't have permission to access /directory/ on this server. Apache/2.2.14 (Ubuntu)Server at localhost Port 80 I have just copied that directory to /var/www. What might be wrong.
View 14 Replies
View Related
May 13, 2010
The company I work for, as usual, is Microsoft-centric. I'm attempting to integrate my Ubuntu server into the domain to allow domain users to authenticate to the server and access file shares using Samba. Here's my current configuration:
[Code].....
View 9 Replies
View Related
Oct 29, 2010
I am trying to configure CUPS and when I click on Adminitrator to Add A Printer, I get Forbidden after entering my root username and password. Anything I'm doing wrong?? In the past, it was just a matter of putting my username and password and walking though the installation. The printer is connected to the computer via USB and is being seen on the machine.
View 2 Replies
View Related
Feb 28, 2011
I don't have permission to access [URL]. I can access [URL] only local. I configured local and network access in /etc/httpd/conf.d/nagios.conf. My access_log: "GET /nagios HTTP/1.1" 403 334 "-" "Mozilla/5.0 (X11;U;Linux i686; en-US; en -US; rv:1.9.1.9) Gecko/2010100330 Fedora/3.5.9-1 fc11 Firefox/3.5.9" "GET /vavicon.ico HTTP/1.1" 404 335 "-" "Mozilla/5.0 X11;U;Linux i686; en-US; en -US; rv:1.9.1.9) Gecko/2010100330 Fedora/3.5.9-1 fc11 Firefox/3.5.9" My error_log: suEXEC mechanism enabled (wrapper: /usr/sbin/suexec)
Digest: genarating secret for digest authentication done mod_python: Creating 4 session mutexes based o 150 max processes and 0 max threads mod_python: using mutex_directory /tmp Apache/2.2.15 (Unix) DAV/2 PHP/5.2.13 mod_python/3.3.1 Python/2.6 mod_ssl/2.2.15 OpenSSL/0.9.8n-fips mod_perl/2.0.4 Perl/v5.10.0 configured --
resuming normal operations Permission denied: access to /nagios denied File does not exist: /var/www/html/favicon.ico Directory index forbidden by Option directive: /var/www/html
View 9 Replies
View Related
Feb 10, 2011
installation moodle on Ubuntu 10.04.1 LTS.at localhost show message It works!at localhost/moodle show message Forbidden You don't have permission to access /moodle on this server.
View 1 Replies
View Related
Sep 21, 2010
Is there anyway i can ssh/rdp/telnet into my server from the outside bypassing comcast ALL blocked ports
View 1 Replies
View Related
Nov 14, 2010
I have installed Redmine according to the instructions on [url] according to the mod_cgi instructions which seem to be the easiest to follow.
The script installed in the installation directory was called 'dispatch.fcgi', and not parsing when I visited the page through the browser, just displaying the contents of the script. OK, so I installed the apache mod libapache2-mod-fastcgi using apt-get. Then 'a2enmod fcgid' followed by an apache restart.
Now when I visit the Redmine folder through the web browser I get 'Forbidden You don't have permission to access /redmine/dispatch.fcgi on this server.'
I checked to ensure www-data owns all files and has correct permissions to the folder. But this doesn't seem to work.
View 3 Replies
View Related
Jun 16, 2011
Using httpd - everything usually works but I have written a cgi script and placed in cgi-bin but i cannot access this dir and keep getting forbidden messages. I have tried adding my root and usual user to the /etc/group file against the user stipulated in the httpd.conf file but i still can't access the cgi script i wrote (in browser - localhost/cgibin/myscrippt.cgi) when i place it in the cgi-bin under htdocs. i have also chmod 755 directory and files.
View 1 Replies
View Related
Jun 7, 2011
I've setup the Uncomplicated Firewall (UFW) on Ubuntu 10.04 LTS and blocked an IP address. UFW status shows that the firewall is active and the IP in question is denied. The issue is that I'm seeing the blocked IP address in my Apache logs.
View 1 Replies
View Related
Jun 18, 2009
If I use my current iptables firewall settings my php includes from my own site get blocked by the firewall. If I use accept all tcp and udp it still does not work. If I use a default policy of accept all, then the php includes work which is not safe. I am running a drupal 6 site on a centos 5.3 vps with apache 2.2.0 and php 5.2.6 w/ safe mode on. He is my current iptable configuration.
Quote:
Chain VZ_FORWARD (1 references)
target prot opt source destination
Chain VZ_INPUT (1 references)
target prot opt source destination
[code]....
View 2 Replies
View Related
Nov 18, 2010
We have a spam in our network and we installed antivirus in all our systems and cleaned the virus from all pc's after that i had removed my ip from the database of blocked ip's but still my ip is blocked for sending spam i don't know from which pc the spam is going on the internet.
so i have a question that my proxy server is redhat linux and as a newbie i don't know the command's to find out which pc is creating large bandwidth to the internet. If you tell the command how to see which pc is sending spam then i will discard that pc. Also i want a strong firewall to stop spam activities.
View 6 Replies
View Related
May 18, 2010
I can't seem to remotely SSH or VNC into my machine. If I'm on the LAN and try accessing via LAN IP, it works fine. If i go in through a remote address (my dyndns) or even my home IP, I can't connect (yes, all of the ports are forwarded, I've triple checked this multiple times). Interestingly enough, port 80 works just fine. It would seem as though some sort of firewall is blocking me. I've done this plenty of times before with various machines, and this has me quite perplexed.
View 9 Replies
View Related
Mar 3, 2010
Basicly just installed a fresh version of Cent OS 5.4 with apache httpd installed automaticly during the installation. The http daemon is running and when navigating to localhost i get the welcome cent os apache page thingy. The problem is when i put an index.html file in /var/http/www then try and navigation to localhost I get a 403 forbidden error.
View 2 Replies
View Related
Jun 9, 2009
I am on mac os x and whenever I make aliases in my apache config file to point to folders outside of main web document root folder, I get access forbidden. I have chmod -R 777 the whole destination directory (to test it) and still no go. Access forbidden. I have also tried pointing directly to [URL].
View 1 Replies
View Related
Feb 18, 2010
In the office there is a local network with samba+openldap PDC. The local domain name is company.net. The company desided to create a corporate Website on a remote hosting and desided that the site's domain should be company.net which is same as local network's domain name. So now it is not possible to reach that corporate website from within the company's local network because, as I guess, bind9 which is installed on above menioned PDC looks for company.net on a local webserver. Is there a possibility to let people from this local network browse the remote site?
View 1 Replies
View Related
Mar 11, 2010
I have installed MoBlock as instructed here: [URL]
After installation I created my own list file in /etc/blockcontrol/custom-blocklist.p2p and have the following uncommented at the bottom of /etc/blockcontrol/blocklists.list:
Code:
locallist /etc/blockcontrol/custom-blocklist.p2p
The list contains the following 2 entries:
Code:
Yahoo:98.137.149.56
Google:74.125.47.147
When I do:
[Code].....
Recently I just noticed that the locallist rules seem to have no effect. I will always get "destination port unreachable" even if the locallist entry in blocklists.list is commented out.
However, whenever I try to browse to that IP, even when blockcontrol is on, even by typing the IP into Konqueror (not the domain name), it lets me go there every time. How can I know that my other applications will not to do the same thing? How can I lock this down and test it empirically to be sure?
View 1 Replies
View Related
Mar 12, 2011
We want apache to server and load our website Only for One specific sub-domain and all other domains should be ignored/dropped or redirect.
View 2 Replies
View Related
Mar 17, 2010
On April 10, 2010, I upgraded some packages on my Ubuntu 9.04 server. This included an upgrade to "ufw 0.27-0ubuntu2". I rebooted the server, and all appeared to be fine.
Now I've noticed that UFW is not logging blocked packets since that reboot. It used to do this. It is still logging the allowed packets that I've configured it to log.
Here's what a "ufw status verbose" says code...
View 2 Replies
View Related
Dec 8, 2010
I come to Ubuntu with the notion that it is much more secure than Windows. In XP I had an anti-virus, third-party firewall and sundry softwares against spybots, rootkits etc. The anitivirus blocked the suspicious web pages while browsing. I generally avoided public networks, carrying a portable internet device Do I need similar stuff with Ubuntu.
View 9 Replies
View Related