Server :: Sqid Resolving Old DNS Entries Of My Local Mail Server
Mar 24, 2010
I have change my mail server previous when i open my domain [url] it goes to [url].Now i have change my Mail server from openwebmail to Zimbra Server .so when i again open to it mail.mydomain.com ,proxy server takes to [url].
If i bypass proxy server then it open mail.mydomain.com(My zimbra Page).i have also updated internal DNS addresss But Squid not updating it
How to update Squid DNS Entry ? how to Update Squid Cache records?
How do I make a local mail server that itself is a client to a WAN mail server.I want the local mail server to query new mail every 30 minutes from the WAN server.
i have successfully done the setup of postfix and dovecot on this link [URL]I run Cent OS on top of window 7 via VM Player,i want to access mailbox from windows 7 using email client like Thunderbird and Outlook. What setting i need to do in my mail server ?
I have a web server that is only accessible to those in the local network. Since it is mainly a testing server, i want to add a test mail server to this system so i can test my programs that email out before i make them public and find out they dont work. I've done this before a while back, but dont recall how i did it or even what the software was(it was for windows, but i want to go linux this time)
i have fedora5 installed on my machine MACH1, which acts as a linux server for few machines. Its over network and we have a dedicated mail server too which is globally accessible. I wanted to set up a local mail server on MACH1, so that i shall be able to send/receive mails from all the users as well as to our other global mail server(on internet).
I installed sendmail in RHEL5.4 with TrendMicro Spam scan engine. The configuration like sendmail should forward all the mails to scan engine after scanning it will deliver it to the mail domains. same way all the mails coming from external servers are scanned and the delivered to local box. My problem is when i sent mail for local users its delivered locally. but when i sending mail for external like yahoo its going through scan engine. I added the smarthost in sendmail.mc file also.
We are using postfix for relaying our internal system generated mail to other domain through google apps. All client servers able to send mail through the postfix server, only user of local postfix server(root, etc) unable to send mail with command like mutt, mail.My detail configuration given below:
I'm trying to configure postfix so smart daemon can send out notifications emails. My system will need to send emails to a mail server that is on the same local network. I have not had much luck in getting this to work, and am confused on what I need to setup to do this and how to set it up. Does my system need to be apart of the same domain as the mail server? If so, how do I do this? Is there any restrictions sending emails from the command line as the root user? Do I need to configure a relayhost?
Basically all I need to do is send emails to a local mail server, i.e. a mail server on the same local network. I do not need to receive emails. I am rather clueless as to how to set this all up, and what I need for this basic email functionality. I have read and tried many postfix configuration tutorials, but when I try to send out a test email to a email address on the local network mail server I always get "Connection refused" to the server and deferred or bounced status on my test email.
Can this be done? A company is currently running a postfix mail server. They got a new server and, while the users are migrated, they need to have the messages delivered on BOTH servers.
I thought it could be easily done in postfix: deliver locally AND forward mail to new server, but all solutions I found would deliver the messages in just ONE place.
how to setup a local mail server for internal testing on my php development work. For example if I sent an email using php script to [url]....I should be able to check the mailbox of 'someone' either in Outlook or SquirrelMail.I have done some reading about this. All I know is that I need Postfix with Courier. But I just don't know to get it working.
This message was created automatically by mail delivery software. A message that you sent could not be delivered to one or more of its recipients. This is a permanent error. The following address(es) failed:
xx@xx.com SMTP error from remote mail server after initial connection: host [URL] [ip address]: 550 Access denied...
I have a problem relate to posttfix.I want to mirgare postfix mail server to exchange 2010 mail server but I can't do it,u can help me.You can show me have to do configure postfix and exchange how to?
There is an requirement, intranet people they may not have internet access but they want to send mail to external domain(internet),but in that intranet network one machine can have internet access. Is there any solution for this requirement.
I've setup my server by following a ton of goods, and it seems to work ok, but I need to start using my server for email in order to receive orders placed via my website. I've followed this guide - [URL] I followed the steps above, and tested the mail server via telnet, and all seemed to be ok. I tried sending an email via Squirrelmail, from cs@thinclientwarehouse.co.uk TO my working email simon@c1systems.co.uk, but the server returned with the following message:
<simon@c1systems.co.uk>: host mail.c1systems.co.uk[95.128.128.129] said: 550-Verification failed for <cs@localhost.thinclientwarehouse.co.uk> 550-The mail server could not deliver mail to cs@localhost.thinclientwarehouse.co.uk. The account or domain may not exist, they may be blacklisted, or missing the proper dns entries. 550 Sender verify failed (in reply to RCPT TO command)
I will be relocating to a permanent residence sometime in the next year or two. I've recently begun thinking about the best way to implement a home-based network. It occurred to me that the most elegant solution might be the use of VM technology to eliminate as much hardware and wiring as possible.My thinking is this: Install a multi-core system and configure it to run several VMs, one each for a firewall, a caching proxy server, a mail server, a web server. Additionally, I would like to run 2-4 VMs as remote (RDP)workstations, using diskless workstations to boot the VMs over powerline ethernet.The latest powerline technology (available later this year) will allow multiple devices on a residential circuit operating at near gigabit speed, just like legacy wired networks.
In theory, the above would allow me to consolidate everything but the disklessworkstations on a single server and eliminate all wired (and wireless) connections except the broadband connection to the Internet and the cabling to the nearest power outlets. It appears technically possible, but I'm not sure about the various virtual connections among VMs. In theory, each VM should be able to communicate with the other as if it was on the same network via the server data bus, but what about setting up firewall zones? Any internal I/O bandwidth bottlenecks? Any other potential "gotchas", caveats, issues? (Other than the obvious requirement of having enough CPU and RAM).Any thoughts or observations welcome, especially if they are from real world experience in a VM environment. BTW--in case you're wondering why I'm posting here, it's because I run Debian on all my workstations/servers (running VirtualBox as a VM for Windows XP on one workstation).
I was using an email server from Icewarp, Merak Mail server V8.0.3 with a keygen for total full free use... I liked this server package... as everything is in there I need and it is controllable from a single window. Now I am planning to move to Linux... They have a Linux version, but not stable and expensive. Is there any thing almost free/less expensive (Specially they need yearly payment for upgrade of even the anti virus and anti spam)?
I need a almost same thing (except of the ftp and VoIP) like Merak (because that I really liked for ease of administration) in Linux? Like a Qmail with real strong Anti Virus and Anti Spam, with unlimited domain and user+autoresponder, mailing list, etc all but controllable from one single GUI?
I have also used external DNS servers just to eliminate there being a problem with 192.168.1.67 (which many other computers are using successfully).
When I attempt to ping Google just to see if it works I receive this:
Code:
If I issue a host command against google.com I receive this:
Code:
I had thought that host and the name lookup routine would use the same DNS server and it should fail on the host command as well as the ping, but evidently not.
I have installed Bind DNS server in Opensuse 11.1 and it was running fine for two weeks. However starting last week it was unable to resolve a lot of internet sites (e.g. opensuse.org). I can say it is working partiallyI tried restart the named service and it worked fine for only about 15 minutes and back to partial condition.All the records in zone files can be resolved. Only some internet websites not able to resolve. The same sites can be resolve with another dns server internally
I have been casually using Unix for about 20 years, but I've never really done anything beyond basic user things unless I had instructions/help.
I've set up an Ubuntu 10.04 system and it works great except for one thing. It doesn't resolve local names.
I can ping or access other systems by IP Address and I can resolve internet names with no problems, but it won't recognize any of the other systems on my home network.
The system is set up for dhcp and everything is defaulted from the installation. There is no DNS suffix or domain set up on my router.
I'm having a strange problem on some of our Debian servers. It all started about three weeks ago when we moved our virtual environment (VMWare ESX3) from a SAN to a NAS (NetApp). At first I thought it had to do with that move but since the other 9 servers are working perfectly I eliminated that idea. For over a year all 12 Debian 5 servers have been working great without mentionable failures. All servers are (where) up to date with the latest patches. About three weeks ago I started having kernel panics with the following message on three of our servers:
Code: Code: Bad EIP value EIP [00000000] 0x0 SS:ESP 0068:f6d7da18
Kernel panic - not syncing: Fatal exception in interrupt and other times it looks like just a dump of hexadecimal data. The only difference between those 3 servers is that they have several mounted shares connecting to the NAS using CIFS. So I was thinking that it might have to do with an update of some kind in regards to smb. I recovered an image from a month ago, before the troubles began, copied over the data and MySQL databases and configured the 'old environment with recent data' exactly the same with MySQL master-master replication, document synchronization and load balancing. This task I performed last night (no other way since it's a production environment). Up to this time neither of the two 'restored' servers had a kernel panic. The one that has not been restored is having one at random about every hour and a half. Following are the different versions between the 'at this time' working server(s) and the failing one:
I've got a machine running Ubuntu Server that is on several VLANs. Each VLAN has its own subnet and the server has an address on each subnet. The switches are set to allow tagged traffic to the server for each VLAN that it is on. Switch ports ending with workstations are given untagged ports on whatever VLAN is appropriate. Workstations are given addresses on a subnet for each VLAN via DHCP. All this works great and hosts on any subnet/VLAN can access the server as normal via its address on that subnet/VLAN.
Accessing the machine by its address on a non-local subnet is where I run into a problem. Inter-subnet traffic has to go through a router, which has been set up appropriately. Running tcpdump on the server and pinging it from a workstation on a subnet, using its address on a different subnet, shows the server receives the ping, but sends no response:
Code: sudo tcpdump -i vlan4 -n tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
I am having problem to collecting email from remote POP3 (all the emails for a domain is stored here) and distribute it after collecting to several users defined to Linux server. I have install postfix in Linux server for email distribution.
I've tried /etc/init.d/mysql start, and it actually starts, or at least, it doesn't error out. But when I do pgrep mysql I get nothing, and when I try to login with mysql -p as root, I get
ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/run/mysqld/mysqld.sock' (2)
None of my other logins work, either, but since pgrep mysql didn't work, it's clear that MySQL just isn't running. I tried uninstalling and reinstalling mysql-server via apt-get, but I get the same thing. The logs in /var/log/syslog show this:
Apr 11 14:31:26 /etc/init.d/mysql[9774]: #007/usr/bin/mysqladmin: connect to server at 'localhost' failed Apr 11 14:31:26 /etc/init.d/mysql[9774]: error: 'Can't connect to local MySQL server
The issue is that my CentOS workstation is in a vlan from where the Intranet's DNS servers are unreachable. For browsing the web there is an ISA proxy server, which I presume resolves DNS for my firefox. However, wget, host, ping and aria2c fail to get any sort of DNS resolution since they're being run from command line.I have exported HTTP_PROXY value, which provides me internet access on console, but,only when I connect using IP address. It fails on name resolution.
My question is:May I redirect the DNS queries to my home PC which would be running a DNS server on a non standard port?I was thinking of putting nameserver 127.0.0.1 in /etc/resolv.conf and then put iptables rule to redirect 127.0.0.1:53 UDP to a.public.ip.address:3535 UDP..I don't know if I am shooting blanks or what, I am not very much aware of this kind of setup.My main need is to provide DNS resolution to console apps.I want to utilize my company's idle bandwidth for bulk downloads, so, using proxy, SSH tunneling through my Home PC is out of question.
I've setup a Lamp Server for Testing, The Lamp Server is Up & Running on CentOs 5.5
I am now trying to setup a VSFTP server where local users can upload files to there home directory so that Apache can serve web pages straight from the directories of system user home/accounts giving users the ability to run their own web sites which are hosted off the main server [tutorial here: [url]
So far i have been able to serve/display index.html files from the users home directory [url] but so far i cant upload files to any user home directory, every time i try to upload a file with filezilla i get this error message: 553 Could not create file. Critical file transfer error
I have searched online for similar problems like mine and so far i've tried alot of the solution but none seem to work. I'm confused, dont know where i went wrong, i put the users in a group called ftpusers and here are the permissions on the users (test, ftpuser & testftp) home directory. have a look an tell me where i went wrong :(
Also the root directory where the web pages are served from is called public_html here are the permissions
Here is my vsftp.conf file can someone check it to see if i made any errors in there: