CentOS 5 Networking :: Benefit Of Creating A Local DNS Server?
Feb 12, 2011
Wondering if the internal network will benefit of connecting to a local DNS server, rather than my ISP dns server. Can I create this local DNS server, without having an external domain, pointing to my server ?
All I want is faster lookup of known hostnames, both internally (hostnames) and externaly (cnn.com etc..)
I am installing a cluster which is hidden from the rest of the world and there is no router either froma general login node. I would therefore require a local repository of the updates to CentOS5.4. When I checked I only found instructions for a full mirror. As I have no room and neither the human resources to setup and maintain a full-blown mirror, this is not a solution. However if I try to find updates for 5.4 to download in a repository-type of way I fail to find a solution. Maybe I am just not looking in the right places.
I'm setting up my server under CentOS 5.5 (text mode) and I installed cherokee server but I couldn't reach it with the ip given by ifconfig. So I figured it might be an problem within cherokee. I also wanted to install webmin so I did that to see if I could reach that one. But also the webmin server is not reachable (I know the port number should be behind the ip ). So if I look up ifconfig I see the server has got an local ip (192.168.1.42) but I can't reach it from another computer in the same local network.
A lot of activity is happening on our local server and I want to know which port is getting the most used and which daemon is using the port. I think it will be some modification of netstat command.
I'm working with a Ubuntu 10.04 LTS system with two network interfaces (both Ethernet). I wish to setup this system such that it is simultaneously connected to my local and an OpenVPN network and able direct traffic between the connections depending on what program is sending the traffic. The problem: Under my current OpenVPN configuration all network traffic is directed to the VPN.
In practice, I would like OpenVPN to operate out of one of my two network interfaces and leave the other interface connected to the local network. Then by default all network traffic should be directed to my local network unless I specify (on a per program bases) that certain traffic should go though the VPN. These two network connections can (should) stay completely independent of each other and do not need to talk to each other.
I've setup a Lamp Server for Testing, The Lamp Server is Up & Running on CentOs 5.5
I am now trying to setup a VSFTP server where local users can upload files to there home directory so that Apache can serve web pages straight from the directories of system user home/accounts giving users the ability to run their own web sites which are hosted off the main server [tutorial here: [url]
So far i have been able to serve/display index.html files from the users home directory [url] but so far i cant upload files to any user home directory, every time i try to upload a file with filezilla i get this error message: 553 Could not create file. Critical file transfer error
I have searched online for similar problems like mine and so far i've tried alot of the solution but none seem to work. I'm confused, dont know where i went wrong, i put the users in a group called ftpusers and here are the permissions on the users (test, ftpuser & testftp) home directory. have a look an tell me where i went wrong :(
Also the root directory where the web pages are served from is called public_html here are the permissions
Here is my vsftp.conf file can someone check it to see if i made any errors in there:
i have successfully done the setup of postfix and dovecot on this link [URL]I run Cent OS on top of window 7 via VM Player,i want to access mailbox from windows 7 using email client like Thunderbird and Outlook. What setting i need to do in my mail server ?
I have created a ftp user in centos 5,but it got all permissions to delete files in other location,view the entire directory and create any folder in every place. How to deny this permissions to the particular user.And please help me to give permissions only to a specified location given by the root.
I had installed PHP using yum install php. I am trying to use the pdf_new function to create pdfs from existing text files, but I get this error PHP Fatal error: Call to undefined function pdf_new()I have noticed that when I run the phpinfo() command, I cannot find the PDF phrase at all. My php.ini file does not even have these two linesextension=php_pdf.dll extension=php_cpdf.dll kind of command I should use if I need to build PDFs using PHP on Centos 5?
I am new to linux, running a brand new centos 5.2 server. One application I want to use it for is to serve as a network host for a game my friends and I enjoy. Normally, to run the game in host mode you call the binary and pass it a port number (along with other options). To host a second instance of the game, same thing different port, you get the idea.
After doing that the binary runs in your window and dumps to stdout, so if you want it to run 24/7 you have to come up with your own strategy like nohup. Fair enough, now, I'm trying to coax the game into restarting automatically upon reboot. The most correct way to do this seemed to be to write a script for init.d so that's the road I traveled down. Now, to strain the metaphor, the pavement has ended and I'm stuck in the sand.
Here begin my questions: I've been following the structure of other init.d scripts and I notice they all seem to call the function daemon() (contained in /etc/init.d/functions) to start their services. Looking at the structure of daemon() I see that you can pass it a user and a pidfile. The user part seems to work fine, but no pidfile is created. Let me be more specific.
Like the other scripts, I explicitly touch /var/lock/subsys/game-port on startup, which works fine. However, all of those other services seem to have a pidfile in /var/run and mine doesn't. They don't create it explicitly in their init.d script therefore I assume that some other process is creating the pidfile. At first I thought it would be the call to daemon(), since you have the option of passing it --pidfile, but that doesn't seem to work.
Are the services themselves creating the pidfile? If that's the case then I have more complications because the game binary apparently doesn't do this. Second question but probably related to the first. None of the other init.d scripts I looked at seem to do anything special to detach their services from a particular terminal session, therefore I didn't think that I would need to either.
Again I thought this was something the call to daemon() might accomplish, but either I'm wrong or I'm doing something wrong. I can probably work around this with nohup or appending '&' or something, but I'm just curious that other services like crond, sshd, named, etc., don't seem to do this. Are they determining this behavior from within the binaries themselves? I hope this is clear, as I said I'm new so I may not be getting all of the terminology correct.
I've been looking for a good tutorial for setting up a BIND DNS server for my local network. What I want to do is..Have BIND running on my home server receiving all DNS requests.Have certain zones (my.zone.lan) pointing to custom IP addresses (I.E. server.lan points to 192.168.{server IP})Zones that don't exist should be passed on to OpenDNS for processing.
We have a squid-proxy configured with 1 delay pool to limit the bandwith to 6M. I have to create a kind of exception for a specific remote host for which we want to reserve 1M which is not included within the 6M. How should I do that? Here is our actual configuration of the delay pools
I've just discovered that crontab is creating a new file in the root directory every time it executes a cronjob, and it doesn't erase over the old file so there are thousands of files in the root directory, they have the same name as the script file (appended with a numeral) but are all blank.here is what one of the cronjob's looks like[URL]
Recently i was able to setup a server which can work as a local yum repocitory in my envornment.However when i tried to update the packages in my repocitory using rsync from mirrors.kernel.org its giving a timeout error. The error i am getting is "rsync error: error in socket IO (code 10) at clientserver.c(107) [receiver=2.6.8]" I am having repocitory for centos 4.4 and 5.2(both 32 and 64 bit versions)
I've just set up the local mirror for my 96 CentOS 5 workstations. The mirroring script is taken from public-mirror howto and looks like this:
#!/bin/sh rsync="/usr/bin/rsync -avHzL --delete --delay-updates" # replaced -q with -v for debuging purposes, removed --bwlimit, added -L to follow symlinks mirror=centos.politechnika.lublin.pl::CentOS # tried several different same result
[Code]....
I tested mirror consistency by putting the mirror addres I'm syncing with, directly into repo file. Than yumex works fine. I tried several different mirrors with the same result.
During the boot of CentOs I see this message which never stops scrolling during all the boot : ERROR 2002 (HY000): Can't connect to local MySQL server through socket '/var/lib/mysql/mysql.sock'
I'm using CentOS 5.5 as a media server to stream media content over my local network. Currently i'm having problem while installing last.fm scrobbler.
Here is a sample output: yum localinstall xmms-scrobbler-0.4.0-7.fc11.i586.rpm Loaded plugins: fastestmirror, priorities Setting up Local Package Process Examining xmms-scrobbler-0.4.0-7.fc11.i586.rpm: xmms-scrobbler-0.4.0-7.fc11.i586 Marking xmms-scrobbler-0.4.0-7.fc11.i586.rpm to be installed Loading mirror speeds from cached hostfile .....
The program package-cleanup is found in the yum-utils package. Upon searching my hard disk this is what i get: find / -mount -name libcurl* /usr/share/man/man3/libcurl-easy.3.gz /usr/share/man/man3/libcurl-tutorial.3.gz /usr/share/man/man3/libcurl-share.3.gz /usr/share/man/man3/libcurl-errors.3.gz /usr/share/man/man3/libcurl-multi.3.gz /usr/share/man/man3/libcurl.3.gz /usr/lib/libcurl.so.3 /usr/lib/libcurl.a /usr/lib/libcurl.so.3.0.0 /usr/lib/pkgconfig/libcurl.pc /usr/lib/libcurl.so
So it tells me that I need libcurl.so.4, but I couldn't find any suitable rpm binaries for it. The only rpm that i found was for Suse and Mandrake. How to set this up, so that i can get a fully functional media server.
I would like to connect via SSH or similar to my servers located in a remote DC from a laptop running centos5. I normally do this on a puter running dows, and using secureCRT. Just wondering if centos has something built in for this, or if there is some preferably free software I can get.
I freshly installed debian lenny using the 5 DVD set that I downloaded from debian.org. I want to create a local repository for all packages that are available in DVD so that I do not insert the disc everytime I install a new software. I have searched various forums but not able to figure the right way to do it.
i have a small issue, to make our network more secure, i now require outgoing email to require authentication. Now the problem..i have a automated mailer that does not have the option to authenticate. is there a way to allow a certain email address or the local network to send out without authentication? If i cannot do this for a single email user to allow them through with authentication, how would i remove the authentication paramaters in the postfix smtp..
I'm in the process of creating local repos for our company servers (CentOS 5.5) and laptops (Fedora 13). And while the CentOS part went perfect the Fedora part is causing major trouble.But first things first, here's the setup: a central CentOS 5.5 server is running Apache2 and has a VirtualHost listening on Port 8080 for both CentOS and Fedora. The DocumentRoot for this VirtualHost is /data/repo wherein two directories, centos and fedora, reside.
This is the .repo-file for CentOS that works like a charm:
Quote:[local] name=CentOS-$releasever - local packages for $basearch
I've got a machine running Ubuntu Server that is on several VLANs. Each VLAN has its own subnet and the server has an address on each subnet. The switches are set to allow tagged traffic to the server for each VLAN that it is on. Switch ports ending with workstations are given untagged ports on whatever VLAN is appropriate. Workstations are given addresses on a subnet for each VLAN via DHCP. All this works great and hosts on any subnet/VLAN can access the server as normal via its address on that subnet/VLAN.
Accessing the machine by its address on a non-local subnet is where I run into a problem. Inter-subnet traffic has to go through a router, which has been set up appropriately. Running tcpdump on the server and pinging it from a workstation on a subnet, using its address on a different subnet, shows the server receives the ping, but sends no response:
Code: sudo tcpdump -i vlan4 -n tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
I just getting back to working with linux. I'm having a problem with what seems to be an iptables thing or something easier than that, Enough here is the question I can get apache to run on the local machine by either going to http://localhost or by the ipaddress on the CentOS machine. But for some reason when I try to take a look on my laptop (on the same network) and try to go to it using the IP address is doesn't go. I can ssh from my laptops but it will not open the apache startup page.
I have recently installed Fedora 12 on a desktop PC and as my first experience of Linux, I am really impressed. I have now installed several packages and have reached a point where I would like to share the PC with other user (family members in the same house).My question seems so basic I am almost embarrassed to ask it but could some one explain the best way to create a local shared directory that could be used to store files accessible to everyone (e.g. music, photos, videos, documents etc.)There will be three users and as it is a family PC, they will all have full access.
Reading posts from various forums, I am little confused about what is the best way to proceed (i.e. what is Linux best practice). The simpler of the two methods is to simply make the directory using the mkdir command, followed by the chmod command to assign full access rights. Fore example if the local shared directory is called 'share'. The alternative approach assigns a group, a group administrator etc and then adds users to the group.
We are in the process for Integration Network with our Government Network. Let say that our network are 192.168.0.0/24 And the Government network are 10.0.0.0/8 I want to know if the local network can resolve the internet names in Government Network. I am using Bind for DNS.