This"407 Proxy Authentication Required ( The ISA Server requires authorization to fulfill the request. Access to the Web Proxy filter is denied. )" is what I get when I try to update or Insatll any thing through Synaptic or atp.But apart from this the Google chrome & firefox works fine.They easily connect to Internet.
I used Avast webfilter (proxied webtraffic through Avast) when running Windows. Sometimes Avast would alert and "protect" me from being infected by a compromised website. NOTE: Avast would alert even absent clicking any links. Just viewing the page could result in infection. Should I be running some kind of proxy webfilter for protection? My understanding is that Firefox can be compromised and this can in turn compromise Ubuntu.Are these kinds of threats specific to Windows running Firefox, or Firefox per se. If Firefox per se it seems like I need some sort of Proxy webfiltering like Avast provides.
In order to have greater control over the emails sent out the company I imagined the following scenario.ScenarioDescribing in words what I need.ser sends an email to user@hotmail.com for example, using the smtp of my isp (is the internet).Before the email is sent, the firewall checks if the recipient is blocked transparent, if you are locked out rejects the email.I searched several forums and mailing lists, but did not find scenarios similar to mine.Where found situations in which the mail servers were internally in the company, which is not my case.
I am getting this error "Unknown filter "police", hence option "rate" is unparsable. This is in Linux Kernel 2.6 which is compiled with the "policing" option enabled in Qos (Kernel config). The error seems to be coming from the file tc.c in iproute2/
static int parse_nofopt(struct filter_util *qu, char *fhandle, int argc, char **argv, struct nlmsghdr *n) { __u32 handle; if (argc) { fprintf(stderr, "Unknown filter "%s", hence option "%s" is unparsable ", qu->id, *argv); return -1; } if (fhandle) { struct tcmsg *t = NLMSG_DATA(n); if (get_u32(&handle, fhandle, 16)) { fprintf(stderr, "Unparsable filter ID "%s" ", fhandle); return -1; } t->tcm_handle = handle; } return 0; }
This is called from get_filter_kind() which is called from tc_filter_modify(), called from do_filter(), from do_cmd() from main()...all in tc.c. Is this a known "bug" in iproute2+2.6 as the 2.4 Kernel works with the same settings?
which is the correct proxy for my company (and works elsewhere) however the add-apt-repository fails with Error reading https://launchpad.net/api/1.0/~freen.../+archive/ppa: <urlopen error [Errno 113] No route to host> This same action works correctly from my home machine that is not behind a proxy. Is there somewhere else i need to specify the proxy for add-apt-repository?
I am attempting to use yum behind a MS Proxy. I am receiving the following error:
#yum list Loaded plugins: fastestmirror Loading mirror speeds from cached hostfile http://mirror.centos.org/centos/5/os/i386/repodata/repomd.xml: [Errno 4] IOError: <urlopen error (-2, 'Name or service not known')>
Trying other mirror. Error: Cannot retrieve repository metadata (repomd.xml) for repository: base. Please verify its path and try again I beleive the yum.conf is configured correctly:
I set up the 'Proxy' from Yast and they tested successfully. I can connect to internet, but why am I getting 'ERROR 407 proxy authentication required'?
FYI, my comp is the part of company .NET network with bunch of firewall and security...
i have just installed Red Hat linux 9 workstation ,i have configured the lan ip and successfully ping to other pcs on network ..but problem is internet is not working , i have already set the proxy setting, i have windows environment and running isa server 2006 proxy server ,when i trying to browse any site , isa eroor page occur "error code :407 proxy authentication required" Linus machine couldnot prompt the user credentiall.
today there was problem with isp but we fix it now other machine run internet fine but when we run through squid proxy machine it give this message to all users. what fields column values config i should check it. i also reset the cache make empty folder. restart machine also service clear the logs . it is on centos 5.4
see below message
ERROR: The requested URL could not be retrieved While trying to retrieve the URL: [URL] The following error was encountered: Read Error
The system returned:
(104) Connection reset by peerAn error condition occurred while reading data from the network. Please retry your request. Your cache administrator is root.
Currently my DHCP Server is working now what i want to have is auto detection of squid proxy in any browser but I still got an error in my dhcp server when I restart it.
My Config:
# DHCP configuration generated by Firestarter ddns-update-style interim; ignore client-updates;
I have a problem with the NTLM proxy we use in school. YAST proxy configuration won't work, it just says authentication error. We have to login with the user in DOMAINuser style and a password, in Firefox it works.
I'm running a squid server on debian lenny with ntlm_auth.
Everything is working good after following those two tutorials : [url] and [url]
As all was working perfectly i decided to rebbot my server, and after reboot the ntlm-auth was not working anymore.
When i tried to run the command "wbinfo" with many options i had this error message "error looking up domain users" or"error looking up domain group" so i think that the problem deals with winbind module.
I know that proxy user needs rights and that after reboot the rights are reseted, so this problem doesn't come from here.
And when i try to open IE with a client i have this "cache accesss denied" because the ntlm_auth wasn't done.
I tried to restart all services (samba,winbind,squid) without any success, only the "kinit" command is still working...
What is going on with winbind when the server is rebooted ?
Code: DEBUG: Executing /usr/sbin/mock-helper yum --installroot /var/lib/mock/centos-5-i386/root install buildsys-build url [Errno 12] Timeout: <urlopen error timed out>Trying other mirror.Error: Cannot retrieve repository metadata (repomd.xml) for repository: groups. Please verify its path and try again at the same time the url works in the FF.
I've been trying to make myself anonymous, but I cant find 'Tor' anywhere, tried 'yum & kpackagekit' neither have it. I did find 'Privoxy', installed it, set proxy for HTTP and HTTPS in Firefox, but it says 'unknown proxy' when I try to use it! I've been to the Privoxy web site and read through the 'User manual', but most of it is 'geek' to me!
I am using Squid Server from last 5years. There is a site "http://www.firstflight.net", which was accessible before few days but now I am unable to access this site. If I use IE8 then getting below error:
"Internet Explorer cannot display the webpage"
Or using Google chrome getting error:
"This webpage is not available The webpage at http://www.firstflight.net/ might be temporarily down or it may have moved permanently to a new web address. Error 330 (net::ERR_CONTENT_DECODING_FAILED): Unknown error."
I am trying to set up my squid3 proxy as a transparent proxy - right now, I have to manually configure browsers to access via proxy. I understand that I have to put some rules into Iptables and also some further directives in the squid.conf.
I have a couple of specific questions. The proxy server is running on a Ubuntu 10.04 workstation and this machine also acts as a dhcp server for the network. I have just one subnet , namely 192.168.0.1-254 There is only 1 network card. Is it much easier to put in a second network card or is it just as easy to configure the existing lan card as a dual IP?
Is it necessary to configure these 2 IP's ( whether they are via 2 lan cards or dual IP on single card ) to be on different subnets. i.e ETH0 192.168.0.1 and ETH1 192.168.1.1 or is ok to have something like ETH0 192.168.0.1 and ETH1 192.168.0.254 ( where ETH0 is the one facing the LAN and ETH1 points to the modem router / switch i.e The Internet ) Where specifically do I save the Iptables rule configuration file and what must I call it ?
I've been doing some security testing in a lab environment that does not have direct internet access. It's actually a little complicated: From home to connect to my lab machine, I
1. SSH to machineA. 2. SSH from machineA to machineB
where machineB is my actual lab machine. neither machineA or machineB allow anything other than SSH, and machineB is only accessible from machineA. However, I really need to run yum on machineB. I have managed to get internet access via Firefox on machineB by creating a series of SOCKS proxy via SSH.
where machineC has internet without limits placed. This is the only way I have managed to get internet working. I tried using ssh -L all the way from machineB->machineA->machineC but it didn't work (even when setting Firefox to use http proxy). I tried using ssh -D all the way, but again that doesn't work either.
I do have access via Firefox using socks proxy. However, yum update fails to retrieve mirror list, and from what I have found I don't believe yum supports socks proxy directly. Instead, it uses http_proxy / ftp_proxy. how to get yum to go out over the SOCKS proxy I created (same one using in Firefox)? It seems like since Firefox can access the internet and everything without issues, i should be able to get yum to tunnel through the same connection to access everything.... I tried
is there a way to force wget to use a specific squid proxy when making connections ? - I use a squid proxy normally, but I need this specific request to go via a different one. I dont have to use wget, I just need a way to test squid's blocking rules by requesting various pages through it, this proxy is not my normally proxy on the network and so I cant rely on wget taking the environment variable.
Also, this is as part of a script, so anything that avoids editing wget config files would be best. - Perhaps curl can do this ? - currently im using the exit code of wget to determine if the connection was made.
I am working on fc10. I connect to internet using two connections: with proxy and without proxy. Initially I had some problem in configuring yum for proxy and I resolved it by creating files proxy.sh and proxy.csh in /etc/profile.d with the required details (export_proxy).Now when I connect to my connection without proxy I have some problems .Whenever I try to do yum update I get the following error:
"[URL]: [Errno 12] Timeout: <urlopen error timed out> Trying other mirror. Error: failure: repodata/primary.xml.gz from adobe-linux-i386: [Errno 256] No more mirrors to try." what changes do I need to make in order to use yum in the connection without proxy?
At the moment I have a proxy and all the users have to configure it in the browser to access internet. I want to make the users able to browse even without configuring the proxy in the browser. but eventually it should be received in the proxy rather than giving an error to the user. I heard with transparent proxy I can redirect all the traffic from a particular network, to a particular host( ie my existing proxy).
I tried this using firewall rules. But then the existing proxy doesn't understand the protocol of the requests. I heard that it should be in the kind of proxy protocol.
I've setup squid proxy st time on centos 5.This is my current setup.squid.conf:Quote:acl our_networks src 192.168.10.0/24 ttp_access allow our_networksQuote:
internet -- modem -- Firewall --switch--squid proxy (192.168.10.100) --client workstation ((192.168.10.200) (client workstation is connected to the same switch as the squid proxy)
I have a problem in my server proxy.I'm using the debian etch distribution, only that now started showing this error:
201.42.212.47 sent an invalid ICMP type 3, code 0 error to a broacast: 0.0.0.0 on eth0 201.42.212.47 sent an invalid ICMP type 3, code 0 error to a broacast: 0.0.0.0 on eth0 201.42.212.47 sent an invalid ICMP type 3, code 0 error to a broacast: 0.0.0.0 on eth0
I want to backup my installed packages using this method mentioned in thread [URL] but the problem is I have 31 broken parkages, i want a way to know the broken parkages so I would remove them in my repack so when i restore the parkages I wont have broken parkages. For now i cant fix them as I dont have internet....
I am new to Ubuntu and would very much appreciate someone's assistance in getting my printer to work in the Ubuntu O.S.The printer filter is located in my /home/ldavis/opt... directory and not in the/opt/ OpenPrinting- Epson/ppds/Epson directory where some of the .ppd.gz are located and where the error message states it is looking for it. Do I need to extract the .ppd.gz in the /opt/OpenPrinting-Epson/ppds/Epson directory and install the filter in the same location?
After trying several times in vain, I attempted to download and install the pinter drivers using the sudo -i apt-get install <> and apt-get upgrade < >. It installed some Epson printer drivers, but I still get an error message that the printer is looking for a printer filter to be installed. Can someone please tell me how to install a printer filter and which directory should it be installed in?
I have a "Forward to" filter set up in Evolution. It halfway works. The emails I want forwarded are matched and put in my Outbox. The problem is that the emails are never sent. I have to manually click the "Send / Receive" button. If left alone, the emails will sit there over the weekend, never sending until I hit Send / Receive.
I need a command that will check the size (visual size, not file size) of every image in a folder and its sub folders, and make a copy (or even better, a hard link) of that file in a second directory if the image is larger than 1920x1080 pixels (in both dimensions, not just total area). Also, lots of these file names have spaces, so the command needs to be space-tolerant
I'm guessing I would need to use one of the imagemagick commands and find, but I'm not sure where to start. I'm still reading man pages, but I thought someone here might save me some time.