General :: Wget Authenticated HTTP Connection Failed
Jul 7, 2010
we have a Red Hat server and I'm using wget in crontab to run some PHP scripts. We've been doing this for some time now and it's been working fine.I tried to add another script using wget to run a PHP script behind HTTP authentication. However, despite the fact that the URL works fine and the username and password are correct, we are getting Connection Timed Out errors each time. What might cause wget to work for unauthenticated URLs, but not authenticated ones?
I've tried --user=/--password=, --http-user=/--http-password and Username:Password@ in the URL and all three fail the same way. Here's the command in question:
Again, wget works, the file with authentication works, but wget calling the file with authentication does not work.
UPDATE: Actually, I get the same timeout if I access the authenticated URL without authentication. Could that mean that Apache is rejecting wget requests for authentication outright?
I would like to find out how I would use both curl and wget to sent an http post to get the hostnames of a few servers. I know am not even given any work of anything I have done, but the reason is that I am really lost, and I do not even know how to start it.
I have a problem setting up a SSH tunnel. I know how it's usually done, but the setup is different this time. I am behind a HTTP(S) and FTP proxy, that does NTLM authentication, and I want to access a server beyond the proxy. MY CLIENT <-> LAN <-> HTTP PROXY <-> INTERNET <-> MY SERVER
So far, the best I have achieved is installing and configuring CNTLM as a local proxy for the authentication part. Using CNTLM, I managed to access and mount a secured (https) DAV share using davfs2. In theory, CNTLM should let me setup permanent tunnels from local ports to distant ports, and it does; however these tunnels don't seem to work for SSH nor for IMAP (another protocol I tried).
I suspect the problem is that neither SSH nor IMAP is HTTP- or FTP-based, but anyway it does not work. So back to square one: how should I proceed to get ssh to connect through the HTTP proxy (with NTLM authentication) to the remote server? For that matter, if there's a better way than SSH to create a tunnel, that would work in my situation, that's OK with me. Just in case, here are the relevant parts from my firewall setup on the server:
# allow continuation of established connections iptables -t filter -A INPUT -m state --state ESTABLISHED,RELATED -j ACCEPT iptables -t filter -A INPUT -f -j ACCEPT
# allow local connections iptables -t filter -A INPUT -i lo -j ACCEPT
# open ports: # ssh iptables -t filter -A INPUT -p tcp --dport 22 -j ACCEPT Yves.
I am having a problem with 5.4 that I did not have with 4.5. The problem happens only sometimes but in specific instances. Basically a summary of the problem is that certain network transactions timeout. The specific instances are with wget, rpm, http. The problem usually, but not always, occurs with pptp stuff. (NOT running pptp but getting pptp stuff). For instance, the following command, which finishes in seconds on non-5.4 OS's: wget [URL] downloads about 20% then gets stuck. About 5 minutes later it downloads another 20% and then gets stuck, etc. The same thing with rpm: rpm -ivh [URL] waits about 3 minutes and then gives an error. I think it does the same thing as the wget but wget will keep trying, while rpm gives up. The error from rpm: Retrieving [URL] ..five minutes later:
I can wget the above as I mentioned before and install it that way. Before I do it, yum works fine. Afterwards, yum exhibits the same behavior of timing out (because it is using the pptp repository). Also visiting the pptp web site from Firefox times out on certain pages. I originally thought it was some problem with the pptp site, but I notice that log into hotmail.com. Does the same thin (fine on other operating systems). A view with Wireshark on the wget (pptp) shows the my machine receiving a reassembled TCPPDU from 216.34.181.96 (Sourceforge), sending an ack, receiving a reassembled PDU, sending an ack, receiving, sending followed by the 5 minutes or whatever of nothing. Then sourceforge sends an RST and a SYN and the process is repeated.
When I put the machine directly on an AT&T IP connection (12.147.X.Y) everything worked fine. Same with Comcast on a direct link. The times I am having problems is when our router is hooked up to a Comcast IP (70.88.X.Y) and assigns 192.168.5.X addresses to our machines. So when I was doing the above from 192.168.5.27 going through the router through Comcast is when I had the problem. So it is probably something with the router, but it is hard to figure out since CentOS 4.5 and Fedora do not exhibit this behavior, nor does 5.4 on most sites (mail.yahoo.com for instance). I did verify, at least from what I could, that ICMP type 3 and 4 are not being blocked. If they were, the same problem would happen on other op systems. And I was able to ping, albeit just locally, but we looked at the router settings and ping was not blocked.
i am using fc9 server i installed Apache web-server i kept some datafile in my html folder when tried to download remotely through web i can download the file tried to get the file in remotely through wget command i am unable to get the fileor is failed: Connection timed out Retrying below the steps i tried it
my target file is http://X.X.X.X/test.zip wget -T 0 http://X.X.X.X/test.zip wget http://X.X.X.X/test.zip
I'm using wget to retrieve a long list of URLs, a small proportion of which fail, hence:
Code: wget --input-file=urls.txt Is there a way to log the urls that have failed? Unfortunatley wget does not output the current URL being processed (and then the status), so hard to see grepping the output helping.
Or should I use some alternative like curl, wmget?
I tried to download some images from google using wget:where wget cbk0.google.com/cbk?output=tile&panoid=2dAJGQJisD1hxp_U0xlokA&zoom =5&x=0&y=0However, I get the following erros:
i am using RHEL5 at VirtualBox,at the time of installation i did create two partition one for / and another for swap.Now my /var directory is taking too much space and i decided to create a new partition with required size and mount that partition to /var.Now http service wont start and i did not reboot yet.
I have a problem with GPG key, when I tried to run #yum updateI have got this errorwarning:rpmts_HdrFromFdno: Header V3 DSA signature: NOKEY, key ID d05c057cGPG key retrieval failed: [Errno 14] HTTP eRROR 404: Not foundI used Centos 5.5.
So we have DNS round robin set up for 4 servers. If we ping dns name (basically an alias) server_connect it resolves with different IP address in round robin format. I.E. x.x.x.1 x.x.x.2 for the 4 different server IP addresses. When we do nslookup server_connect it will come back first time as server1_connect server2_connect through server4_connect so the server is able to resolve through ping and nslookup resolving the initial dns name (alias) to the dns name associated in the round robin. Problem is when we try to connect with http or telnet it comes back host unrecognized. I can put one of the 4 round robin servers in /etc/hosts and it connects fine so I'm thinking that either one of three things. 1) ttl 2) It does double connection first to identify itself to the round robin server and then handshake but second time it hits for the handshake the IP and dns name is different than what it expected so it fails. 3) Since we are trying to telnet to dns alias and it is returning different dns name it fails.
2 and 3 seem most promising but now I'm at a stand still.Anyone else come across this issue and if so how did you resolve.
My Phone keeps reporting Cannot Get Mail The connection to the server failed My phone does seem to show (all?) the contents of my home directory so must have received some data but no actual mail.
In debug.txt, this keeps repeating over and over: Found 0 message changes Initializing message diff engine 1 messages in state
And in /var/log/maillog, this does: Killed (lost mailbox lock) There are no other imap connections running.
My problem is the following:I can only connect to the internet and download through the FTP, but not through HTTP.I am in a hospital at this moment and have a netbook (Lemote Yeelong). Also there is a computer here with internet access.I would like to access the internet by using the cable from the computer.I configured my netbook just like the local computer (same host name, same MAC address, DHCP), just to be sure.I can ping google.com and I also can download from ftp.debian.org though ftp.However, when I try to connect to google.com through epiphany, it just loads and loads without success
I'm trying to open a web page to automate some data checking, and I'm getting a "too many redirects" exception. I'm not experienced enough with Java to know what to try next, and would like some help.
I'll show the code for what I've tried, but first, details about the website:[URL].. The website is a series of aspx pages. Going to the above address in a browser gets you to "default.aspx"--a search page. You don't see the redirects (specifically to a Login.aspx page--which redirects back to default.aspx). I can get the search page with either Firefox or wget--and on the same computer that I'm writing the Java code on. So, Firefox and wget are doing something that my code isn't. Also, I'm not behind a proxy.
I've searched about the redirect exception, and one or two pages blame poor website development. That may be true, but I don't have any control over the website. So, fixing the website is not an option.
I did find this page where someone has the same issue. I used some code given in one of the responses to discover the default.aspx->Login.aspx->default.aspx redirection loop.
Much of my code was pulled from the Working with URLs and Working with Cookies tutorials from Oracle. I tried adding the cookie handler thinking that maybe Login.aspx was trying to create a session ID or some other connection-specific identifier. But either cookies are not the solution OR I just didn't code it the right way.
[Code]...
After more searching, I'm 99% certain it's a cookie-handling issue. I added some more code (not included in the above) that examines the full response from the redirect to the Login.aspx page. The response includes a Set-Cookie header for "ASP.NET_SessionId". Now to find some code that will store and send the session ID.
I installed Linux 6, configured the network during the install with a static IP and when I do an ifconfig command it doesn't show an IP Address. Can't get out to the Internet because I don't think I have an IP/Address. I restarted network, went into preferences-network connections and everything looks fine, ran system-config-network and everything looks fine. When I do an ipup ifcfg-eth0 and get the message "Connection activation failed: Device not managed by NetworkManager". Is there an option I failed to load at install or I need to install now?
What is the simplest way I can get a secure, authenticated SMTP server running on my Linux machine so my family can use it to send mail from hostile networks?
Currently I have exim bound to 127.1:25 so if I can send mail from my laptop using ssh -L2525:localhost:25 wolever.net... But I'd like exim (or some other SMTP server) to bind to, eg, 0.0.0.0:2525 and accept authenticated SMTP connections.
I am trying to setup multipath with failover policy on openSuSE 11. I have two qla2xxx HBA's installed and they appear to be working. Here is the output of "multipath -l" command
[Code]....
While testing, I pulled one of the two connection to SAN, and the connection failed over to second HBA connection to SAN. When I plug the cable back in, it does not fall back to original connection... It stays in failed state. Also, I noticed that failed disk (sdd disk) comes back as (sdg disk), which is probably why connection does not fall back to original HBA. But, when I run "/sbin/service multipathd restart" sdg disk shows as as enabled in multipath -l...
I tried to finally set-up surround sound on my Ubuntu 10.04 machine and i was successful in doing so after a while of changing settings and tinkering. After doing so though, i was later in the day not able to get into the Pulseaudio Volume Control, therefore, everything is stuck where it is and i can't change anything. When i try to open Pulseaudio Volume Control i get the error message, Connection failed, Connection refused. My programs still play sound and i can control the sound of the programs through the programs, but i cant control the sound of the sorround sound now and my main sound bar only affects my stereo speakers, not the surround ones.
I believe this may have happened after the random disk check on restart, but i can't 100% confirm that and i'm not even sure if they are related. I still have sound, but i have no way to control the volume of my entire system at once. I did some looking around online and saw that others are having or had the same issue, and it was through different versions of Linux also. somehow reset everything so i can access PAVC and start over . I also tried steps listed in another thread here on the same topic, but they weren't working and it also seemed to be for older Ubuntu versions.
I got an error while working with NetworkManager in new RHEL6/ Fedora14. I have a ZTE AC 2726 3G wireless modem. i had tried to connect it via NetworkManager . It only requires username and password to connect but the thing is that after putting all the credentials i had checked the option" available to all the user "and while i was trying to save the profile it shows the following error saving connection failed 0 can't write connection type 'cdma'I had google so many times around this error but no better solution is there .
vsftpd not working on my CentOS system. I tried logging in using FileZilla but the error message was: Connection attempt failed with "ECONNREFUSED - Connection refused by server".
Iam using socket programming in one of my problem. The scenario is like this.
=>one module is working as a socket server(process p1) is able to handle client sockets on that port. =>one module is working as socket client (process p2) is connected to server socket and Tx/Rx data on this socket. This module has some more threads, based on received data from server socket it will connect to http on other thread and get information... =>one more module is working as socket client (process p3) is also connected to server socket and doing some other transactions.
Here my problem is when in presence of p2, p3 http connection taking more time than the expected. If we wont start p3(means only one server-client socket) then http connection is fast enough.
I have shared keys setup on my domain, so I never type my password to login anymore.
I've forgotten my password now. This is a problem because only my user can sudo. Password authentication for root has been disabled, so without my password, I cannot do maintenance on my web server.
Is there a way to reset my password as my [now only] key-authenticated user?
I am trying to create a certificate case user logon via ssh. On the server I have openSSH and a few users. I want to be able to assign a user a certificate to connect remotely via SSH.
If a wget download is interrupted (like if I have to shutdown prematurely), I get a wget.log with the partial download. How can I later resume the download using the data in wget.log? I have searched high and low (including wget manual) and cannot find how to do this. Is it so obvious that I did not see it? The wget -c option with the wget.log as the argument to the -c option does not work. What I do do is open the wget.log and copy the URL and then paste it into the command line and do another wget. This works but the download is started from the beginning, which means nothing in the wget.log is used.
I need to install a program by using the address http://255.255.255.255. However, when I type this address in my browser, I get the following error: "Failed to connect. Firefox can't establish a connection to the server at 255.255.255.255. Though the site seems valid, the browser was unable to establish a connection." Is there an easy way to put this site into the air?
I am trying to backup a client called greetings.com, but when I do an amcheck on the server, I get: WARNING: greetings.com: selfcheck request failed: Connection refused
I got this error for two days.back then I googled this error and i got some information about log file in /var/log/syslog but i couldn't find it in google at all. this is all i got from syslog . look at the red paragraph
Using netcat, nc(1), craft a valid http/1.1 request for getting http headers (not the html file itself!) for the main index page of www dot aalto dot fi. What request method did you use? Which headers did you need to send to the server? What was the status code for the request? Which headers did the server return? Explain the purpose of each header.
nc -v www dot aalto dot fi 8080 HEAD / HTML/1.1 host: www dot aalto dot fi And it returns: 200 OK Content-Length: 858 Content-Type: text/html Last-Modified: Thu, 02 Sep 2010 12:46:01 GMT [Code]....
I really don't know what does it mean. Question 2: Using netcat, nc(1), start a bogus web server listening on the loopback interface port 8080. Verify with netstat(, that the server really is listening where it should be. Direct your browser to the bogus server and capture the User-Agent: header "Direct your browser to the bogus server and capture the User-Agent: header" I don't understand this question.
I installed Nagios on my Ubuntu 10.04 server using apt-get and when I accessed the web console, everything was OK. I made some changes to apache (creating some new virtual sites) and since then Nagios gives me a warning message for HTTP with the message, HTTP WARNING: HTTP/1.1 404 Not Found. The sites that I created are working perfectly. I noticed that the attemps are 4/4. Does this need to be reset or does Nagios automatically reset that once it detects the issue is resolved?