We've had this server running before but two hours ago there are some domains outside which we cannot send email to. this is a sample of the message we get
Quote:
The original message was received at Fri, 22 Jan 2010 14:31:59 +0800 from [10.0.200.10]
----- The following addresses had permanent fatal errors -----
<beth@nippon-antennaphil.com> (reason: 550 localhost.localdomain does not exist)
----- Transcript of session follows -----
... while talking to nippon-antennaphil.com.:
>>> DATA <<< 550 localhost.localdomain does not exist 550 5.1.1 <beth@nippon-antennaphil.com>... User unknown 451 4.4.1 reply: read error from nippon-antennaphil.com.
Our server's been running smoothly before but i dunno what happened. I can send mails to gmail and yahoo as well... but there are like 3 domains which we cannot send mails to.
I have been beating my head for the last few weeks on this problem, (although I have been taking the wrong approach, it seems).
I need a gateway to direct web traffic to three separate servers/domains. I have been trying to do this with both a dns server and , (seperatly), apache server to forward requests. The dns server was a no go, and <i can only get apache to redirect http and ftp.
After Googling this ALOT, I believe that what I need is a gateway server to redirect my traffic to the 3 different servers. I have been reading about using using nat and iptables for this and was wondering if anyone had any advice/suggestions on this. The other thought I had was to use something like pfSense to create the gateway, but I am still reading the documentation, and I am unsure if this approach will work.
I've been attempting to setup PXE/HTTP network installs so we can better handle deployments for new systems. I have a test CentOS 5.4 VM running, and another test VM that I want to deploy 5.4 to. TFTP, DHCP are all working correctly. Apache 2.2.3 config "seems" OK. When I kick off the VM which I will install to, DHCP discovery and IP allocation works, the TFTP server is found, I am presented with a menu option of OS selection.
I choose #1, for my 5.4 but then it immediately tells me: "Invalid or corrupt kernel image"
/var/log/messages doesnt show anything other than the DHCP OFFER/ACK process and that TFTP client doesn't accept options /var/log/httpd/error_log doesnt show anything either Not sure where else to look for diagnosis. My Apache config directory: /var/www/html/CentOS
My Apache DocumentRoot: /var/www/html/CentOS Directives: <Directory "/var/www/html/CentOS"> AllowOverride None Options None Order allow,deny Allow from all </Directory>
Forgot to add this line from my pxelinux.cfg/default file: label 1 kernel 5.4/vmlinuz append initrd=5.4/initrd.img ramdisk_size=16000 method=http:/10.37.129.3/CentOS ip=dhcp
(I think I have found my problem.. 5.4 was in the TFTPBOOT directory, but now that I'm using HTTP, I changed this to be: kernel CentOS/vmlinuz and append initrd=CentOS/initrd.img) -- question is, will just changing this work?
My ISP has a bad habit of killing nearly all sustained FTP transfers (reset packet to my router and the remote host) and I need to install CentOS 5.5 on an older pc that doesn't have a DVD drive. My solution would be to (temporarily) turn on IIS on my laptop and extract the contents of the DVD to the ftp dir, but do I simply just drop everything in there with the same directory structure as the DVD or is there a particular way I should set up the directories?
I'm trying to set up a server for an NPO who connects to the Internet through an internal proxy (Websense). I can't access the Internet unless through the proxy.
Since it's a server I did not install a GUI so I'm wondering how to connect the new server to said proxy server. I have the IP address, port number and username-password for the proxy server. The new server has a static IP address and is ready too go!
i am forwarding HTTP request to a internal server, it is quiet successful but access logs donot show the ip of the external m/c. Rather it shows the ip of the machine on which i have enabled port forwarding.
I installed Nagios on my Ubuntu 10.04 server using apt-get and when I accessed the web console, everything was OK. I made some changes to apache (creating some new virtual sites) and since then Nagios gives me a warning message for HTTP with the message, HTTP WARNING: HTTP/1.1 404 Not Found. The sites that I created are working perfectly. I noticed that the attemps are 4/4. Does this need to be reset or does Nagios automatically reset that once it detects the issue is resolved?
Some time yesterday, I lost the ability to ssh my remote server, or even visit any webpages it hosts.
I've explored hosts.deny, /var/log/secure and even turned iptables off to see if it would fix anything. To no avail. Here's what my ssh login attempt looks like:
I'm trying to setup a Centos 5.6 Squid Proxy Server with Content Filtering & Antivirus Scanning Incoming HTTP Traffic from the Internet
I then proceeded to setup an configure the Proxy Server, i was able to test and confirm that Squid and Dansguardian Content Filter is working, however i dont know if Clamav is scanning HTTP traffic before it hits the client/server. Is there a way i can check if the Antivirus scanning is working.. is there some log file or real world test i can i can do to confirm that Clamav is scanning incoming traffic or even blocking potential viruses ??
Anyone who has squid proxy server with Clamav configured and its working can share there settings/setup with me and how they tested it ??
I use sendmail using my server. I hosted my site I just bought a domain name. Everytime I send mail I saw in my log files localhost.localdomain. How can I change this hostname in sending my mail.
here's the log:
From MAILER-DAEMON@localhost.localdomain Fri May 8 16:30:23 2009 Return-Path: <MAILER-DAEMON@localhost.localdomain> Received: from localhost (localhost) by localhost.localdomain (8.13.1/8.13.1) id n488UNlD006632;
how do i able to allow some users that are able to create content in directory of http server. For example: i have configured a web server which have default document root /var/www/html, now i want to extend my web server through virtual hosting , i have enable virtual hosting, but i want that user sumit is able to create content in /var/www/html/secret. which is the document root for my virtual site?
In my workplace, I can't update Centos using http protocol, but, it's possible with ftp (I tried with Debian and it's work). How to configure the repositories in Centos 5.5, to download the packages by FTP?
I have a debian box running Apache2 and PHP5.2.6 lenny.
When a request is made via https, php displays the content fine. If the request is made over HTTP the file is offered for download, rather than displaying it.
I know its probably something trivial but I've never seen this issue.
The plot thickens, I can display PHP over HTTP in some directories but not others (which offer the file for download)?
Squid document says that Squid accepts only HTTP requests but speaks FTP on the server side when FTP object are requested.
We call Squid HTTP and FTP caching proxy server. Does it also caches FTP contents? Is it possible to configure FTP clients to use Squid cache? When we make an FTP request to an FTP site via Squid will it be bypassed?
When I attempt to install a new hardware system over the wire via NFS I get the following error:"unable ro retrieve http://192.168.0.1//install/images/stage2.img."The httpd.conf file is configured to use /var/www/html as the root. I have placed the installation files in /var/www/html/install. In the http install screen I enter the following information:Web site name: 192.168.0.101CentOS directory: installI have tried many different variations in these two entires, but the error always is the samw: the installation script searches for an address which has an extra / character interposed between the server IP address and the directory path entered in the next line.
Using netcat, nc(1), craft a valid http/1.1 request for getting http headers (not the html file itself!) for the main index page of www dot aalto dot fi. What request method did you use? Which headers did you need to send to the server? What was the status code for the request? Which headers did the server return? Explain the purpose of each header.
nc -v www dot aalto dot fi 8080 HEAD / HTML/1.1 host: www dot aalto dot fi And it returns: 200 OK Content-Length: 858 Content-Type: text/html Last-Modified: Thu, 02 Sep 2010 12:46:01 GMT [Code]....
I really don't know what does it mean. Question 2: Using netcat, nc(1), start a bogus web server listening on the loopback interface port 8080. Verify with netstat(, that the server really is listening where it should be. Direct your browser to the bogus server and capture the User-Agent: header "Direct your browser to the bogus server and capture the User-Agent: header" I don't understand this question.
I setting up a web server on my Linux (Centos) using "Apache" web server. And its working well, it will show my websites. But when i try to put my url in a internet browser (ie. only type "myweb.com" ) is only be [URL]. Usually as i know like the other websites (ie. google.com) it will go to auto replace the name be [URL]. But my url here it's not be replacing like that. How to do this configuration. I don't know where the services that i need to look. (ie; named (bind) or in Apache web server it self)?
I am having a problem with 5.4 that I did not have with 4.5. The problem happens only sometimes but in specific instances. Basically a summary of the problem is that certain network transactions timeout. The specific instances are with wget, rpm, http. The problem usually, but not always, occurs with pptp stuff. (NOT running pptp but getting pptp stuff). For instance, the following command, which finishes in seconds on non-5.4 OS's: wget [URL] downloads about 20% then gets stuck. About 5 minutes later it downloads another 20% and then gets stuck, etc. The same thing with rpm: rpm -ivh [URL] waits about 3 minutes and then gives an error. I think it does the same thing as the wget but wget will keep trying, while rpm gives up. The error from rpm: Retrieving [URL] ..five minutes later:
I can wget the above as I mentioned before and install it that way. Before I do it, yum works fine. Afterwards, yum exhibits the same behavior of timing out (because it is using the pptp repository). Also visiting the pptp web site from Firefox times out on certain pages. I originally thought it was some problem with the pptp site, but I notice that log into hotmail.com. Does the same thin (fine on other operating systems). A view with Wireshark on the wget (pptp) shows the my machine receiving a reassembled TCPPDU from 216.34.181.96 (Sourceforge), sending an ack, receiving a reassembled PDU, sending an ack, receiving, sending followed by the 5 minutes or whatever of nothing. Then sourceforge sends an RST and a SYN and the process is repeated.
When I put the machine directly on an AT&T IP connection (12.147.X.Y) everything worked fine. Same with Comcast on a direct link. The times I am having problems is when our router is hooked up to a Comcast IP (70.88.X.Y) and assigns 192.168.5.X addresses to our machines. So when I was doing the above from 192.168.5.27 going through the router through Comcast is when I had the problem. So it is probably something with the router, but it is hard to figure out since CentOS 4.5 and Fedora do not exhibit this behavior, nor does 5.4 on most sites (mail.yahoo.com for instance). I did verify, at least from what I could, that ICMP type 3 and 4 are not being blocked. If they were, the same problem would happen on other op systems. And I was able to ping, albeit just locally, but we looked at the router settings and ping was not blocked.
I have a problem with GPG key, when I tried to run #yum updateI have got this errorwarning:rpmts_HdrFromFdno: Header V3 DSA signature: NOKEY, key ID d05c057cGPG key retrieval failed: [Errno 14] HTTP eRROR 404: Not foundI used Centos 5.5.
I have a problem installing CentOS 5.5 over lan. I set up a dhcp3 server and it works, client got an ip. I set up a tftpd-hpa server and it works, client can boot the install image of CentOS 5.5 I set up an apache server and it seems to be impossible... In the installation marathon I clicked through the menu to the choice of the web server where my installation files are available. I enter the ip 192.168.138.1 and the path CentOS-5.5 and CentOS tells me "Unable to retrieve http://192.168.138.1//CentOS-5.5/images/stage2.img" . Thats right because the two slashes are not right. Where the hell come the two slashes from???
I have some photos posted in [URL] Into each caption I've added a link to my server to let friends download them in larger sizes. tail -f access_log only displays some of those accesses, I don't understand why. If I reload large image page an input is recorded and displayed from access_log What can be happend?
I'm working on an application that requires a large amount of storage space and I want to handle storage `in-house` (Much cheaper than, say, S3) so we will have multiple servers (Initially 4) with large amounts of storage (6TB each). The storage will need to be very flexible and configurable, each piece of data should be replicated on at least 2 servers and must be easily readable/writable from ether an API of a UNIX device/file/folder like a normal drive, I don't mind which. We must also be able to easily offload content to our HTTP CDN (Edgecast), it doesn't need to have built in HTTP support but if it doesn't I'm going to have to write something to get the files onto HTTP so they can be pulled by the CDN.
I've looked at a lot of solutions including Eucalyptus Walrus OpenStack Object Storage MogileFS MongoDB GridFS (I'm not sure why, it just sounded cool =) ) and some others which I can't remember
All the servers will be running RHEL 6, they have 4x1.5TB drives which will be RAID1'd into a single partition. All the servers have 1GB/s connections between them and 100MB/s connections to the internet with unlimited bandwidth. They have 2x2.66ghz processors. I understand there isn't a single, perfect answer but it would be nice to get some pointers.
Been a while but have a few scripts that need to hit a website that's local to that network, but also a public site. Currently there is an .htaccess in that folder with this lockdown;