Ubuntu Servers :: Bind Fails First Attempt, Then Works?
Jan 10, 2011
I'm having an issue with a BIND server. After a restart, (or randomly, I assume whenever a cache expires,) when I try to resolve any domain I get a "Host yahoo.com not found: 2(SERVFAIL)" Eventually it starts working and works fine till the cache expires again;
The nm-applet starts automatically as it should, and also tries to connect to the local, wireless LAN, but always fails its first attempt. If I click the icon and ask nm-applet to connect again, it always succeeds.
The server, running the wireless LAN, never sees the first attempt to connect, so no wonder this attempt fails. But it would be nice to make it succeed!
I'm running F12 and Package Updater has been telling me to update Thunderbird for two days but fails on every attempt. I tried to go the yum route today with yum remove thunderbird yum install thunderbird
And I get this error: Error: Package: thunderbird-3.0.10-1.fc12.i686 (updates) Requires: nss >= 3.12.8 Installed: nss-3.12.7-6.fc12.i686 (@updates) nss = 3.12.7-6.fc12 Available: nss-3.12.4-14.fc12.i686 (fedora) nss = 3.12.4-14.fc12
Is there any way I can update nss such that Thunderbird will have access to the correct version or is this something that is in process for the repository?
I am dual-booting 11.04 alongside windows 7. I shrunk my w7 partition, and would like to extend my ubuntu partition to fill up the remaining space. When I boot from GParted live cd, and attempt to 'move/resize' my ubuntu partition, it simply fails. It doesn't really give an error message either, simply 'failed to move/resize [partition name]'
I am attempting to install debian for the first time on my pc that has no cdrom drive. I downloaded the Jessie CD image and wrote it to a 4GB stick, it didnt work. Then tried the netinstall image but face the same issue.
To write the usb stick I used unetbooting first, then tried win32diskimager and finally tried DD while stick was not mounted
Issue is still the same: I boot from the usb and after selecting language and keyboard it fails to detect cdrom drive (no drive at all in my pc). Same behavior using normal or expert mode.
Im also unable to manually specify the drive (it looks at /cdrom and I wanted to change it to the usb stick itself or mount usb to /cdrom but I cant find my stick in /dev)
I have updated bind using yum on a Centos 5.3 server, after restarting, I have this error now. Error in named configuration: zone localhost/IN: loaded serial 42 /var/named/mydomain.hosts:20: unknown RR type 'SPF' My version of is : bind.x86_64 30:9.3.6-16.P1.el5
So i have a fresh install of the server edition of Karmic, i'm running the Xfce desktop. When I attempt to manage users and groups through the GUI, I am prompted for what I think is the root password, the reason I say this is because the account I am currently logged in has sudo privileges and it does not accept that password at all, but I read that by default the root account is 'locked,' (to be honest it was so long ago since I last installed Ubuntu I completely forgot if it is or isn't, my current desktop installation has su access) is it asking for the root password? why doesn't my current user account password work if the root account is 'locked'? I can perform all other administrative tasks with sudo no problem.
the funny thing is, I have the exact same setup in a virtual machine, the same problem happens, except for some strange reason after changing the password on the only account (besides root), the password required to administer users and groups stayed the same after the change. (at the time of installation I just put both the user and root password the same and now that it is setup), i'm now ready to change the passwords. except now I read that the root account is locked by default, but this strange problem occurs.
I'm trying to use a remote procedure call. When I call my server, my server should activate gammu and send an sms with it. I've used the code in the following tutorial: [URL]. The command for the uptime, and the command for the greeting work perfectly. But when I write my own method on my server, it fails ..
Code: function uptime_func($method_name, $params, $app_data) { return `uptime`; } function greeting_func($method_name, $params, $app_data) { $name = $params[0]; return "Hello, $name. How are you today?"; } function gammu_func($method_name, $params, $app_data) { $text = $params[0]; $number = $params[1]; $result = "echo '$text' | gammu sendsms TEXT $number"; exec("$result"); return $result; } On my Client (the one that calls the server) I see the output of $result. So my gammu_func is definitely working... He just doesn't execute Code: exec("$result")
I know that syntax is right cause I tried it in a different php file. I think it has something to do with the user rights. I don't think I have the privileges to run that command ...
I want log in locally to my Lucid (10.04) workstation and have my code saved over the network on my samba account At work, all developers have samba user ids and when we were running Red Hat, we went thru the following procedure to get setup.
* open a shell session to NFS server and execute the "id" command to get my samba user information.
Code:
$> ssh l alberto our_nfs_server $> id uid=7090(alberto) gid=100(users) groups=100(users)
* Create locally a login on my Linux workstation with the same login and uid.
Code:
$>sudo useradd u 7090 g 100 d /home/alberto s /bin/bash alberto
[code].....
Code:
mount error(13): Permission denied Refer to the mount.cifs(8) manual page (e.g. man mount.cifs) mount.nfs: access denied by server while mounting our_nfs_server:/d0/homedirs/alberto It does not work at all!!?? However, the same procedure on a RedHat workstation works fine. No errors.
I was wondering if anyone knows how to make a zone file for bind dns server that points "example.com" and all (by all i mean any) sub-domains to this 127.0.0.1 address. and i wanted to set up a virtual host in apache to use /var/www/eample/index.php for 'example.com' and all of its sub-domains.
I am able to boot from the Ubuntu 10.04 LiveCD fine, but the Kubuntu CD fails to boot. Trying to run Kubuntu sends me to a command line and after the first three steps of the install option I am again sent to a command the line. I have burnt and verified several Kubuntu CDs and each time I get this result. I am running Ubuntu, but I would like a fresh install of Kubuntu (I would rather not just install kubuntu-desktop).
I have a usb modem which I plug into my ubuntu 10.10 system for a dial-up service. The modem is recognised by wvdial and dials properly, but the connection is not established.I get "Unable to run /usr/sbin/pppd" (although pppd is certainly in /usr/sbin), followed by "Check permissions of specify a 'PPPD Path' option in wvdial.conf", and finally "Connected, but carrier lost", then it gives up.
Have basic LAMP setup on 9.10 box. I want to have a publicly accessible website AND I want to have phpmyadmin available. The only thing is I would rather not have the phpmyadmin interface available on the internet. I usually open a ssh port forwarded tunnel when I need to use phpmyadmin on this server. I want to add a directive to make phpmyadmin bind only on localhost. I have found the phpmyadmin config file in /etc/apache2/conf.d
phpmyadmin.conf -> ../../phpmyadmin/apache.conf
I have tried adding some LISTEN directives, but apache does not like my directives-- I am obviously not doing it right. I have looked for a bit on the internet and can't find out how to disable external access to a configured site in Apache.
I have searched some about Bind9 and have it installed in my server. Now I'm looking for a Tools like mysqlBind and I found unxsBind is the next version of mysqlBind (These tools help you config BIND DNS with MySQL). But the issue is: I don't know how to Install it in Ubuntu - there just say how to install with Centos.
I need a bind-server (I mean DNS server) with a SQL backend. As far as I have seen, the only viable option I have is to recompile bind9 with support for either MySQL or PostGre.
Link to this: [URL]with an additional
Code: apt-get install libpqxx3-dev
From some tutorials I have read, I see I need to recompile it to enable its SQL support.
But when I do it that way, I won't get any security updates, won't I?
So... Is there any bind server with SQL support in the repositories ? For that I don't have to worry about security updates...
System:Tri-head, dual-card: GeForce 9500 GT, GeForce 8400 GS Dual-boot: openSUSE 11.4, Ubuntu Natty Driver: nvidia proprietary (260.19.44 in openSUSE, 270.30 in Ubuntu due to kernel version) xorg.conf: same for both Results: All three heads work just fine in Natty; secondary screen fails in openSUSE:
Code: [25.164] (EE) NVIDIA(1): Failed to initialize the NVIDIA GPU at PCI:2:0:0. Please [25.164] (EE) NVIDIA(1): check your system's kernel log for additional error [25.164] (EE) NVIDIA(1): messages and refer to Chapter 8: Common Problems in the [25.164] (EE) NVIDIA(1): README for additional information. [25.164] (EE) NVIDIA(1): Failed to initialize the NVIDIA graphics device! [25.164] (II) UnloadModule: "nvidia" [25.164] (II) UnloadModule: "wfb" [25.164] (II) UnloadModule: "fb"
But nothing is jumping out at me in the output of dmesg. I also don't see any additional system or kernel logs in /var/log. I'll google some more on that front. One other fun fact: nvidia-settings fails to run in openSUSE. Unless I launch it under gdb. Then it starts up and runs as expected. (And the second screen ain't there, as expected.) Here's (what I think are) the relevant items: Xorg.0.log - Pastebin.com dmesg - Pastebin.com xorg.conf - Pastebin.com Additional output available upon request.
After a massive meltdown with the upgrade, I finally got the system put back together. Had to reinstall grub. It now defaults to a version that says "Generic - pae". When choosing this from the grub menu, it always hangs on the Ubuntu splash screen; however, I found that when I choose "previous Linux versions" from the grub menu, then choose the Ubuntu-generic, everything works fine - other than my initial bad impression of unity - but it works, at least.
why my default choice in grub doesn't work, and what I need to do to fix it? Below is a cut and paste from /boot/grub/grub.cfg. The first entry is the one that does NOT work, and the first entry after "submenu" is the entry that DOES work.
menuentry 'Ubuntu, with Linux 2.6.38-8-generic-pae (recovery mode)' --class ubuntu --class gnu-linux --class gnu --class os { recordfail set gfxpayload=$linux_gfx_mode
I have 2 nameservers setup, a master and a slave. when I first setup the slave, I restarted bind9 and all of the zone records propagated just fine. Today, I updated one of the records on the master (no problems), but when I restarted bind9 on the slave it gave me a FAILED message. I checked the log and it was trying to receive notify's from my gateway address (192.168.10.1), and got "Failed to update from non-master". I did some research and found several people having a similar problem, but their slaves were trying to update from their own IP, not from the gateway IP. I tried their solution (allow-notify { 192.168.10.1; }) but all that did was allow the slave to restart bind without errors, it still doesn't update the records. It says "zone is up to date" but it's not.
However I just built a 10.10 server, installed webmin, vmware, and the server is working perfectly. I configured my bind 9 server using the latest webmin and on the server everything resolves perfectly to both the internet and lan. I have it set to 127.0.0.1, the server ip address is 10.1.50.25. However, it will not accept dns client queries in which they cannot resolve to the lan or internet. I have the dhcp giving out the dns server 10.1.50.25. NSLOOKUPS from the client show query refused. I know there has to be some setting or config that will allow clients to query but I am not able to locate it, and I am not totally knowledgeable of named.conf and been all through the webmin module and configuration settings.
I'm trying to install Mandriva Spring 2010 for a friend on his laptop after MS Windows crashed. The installation appeared to work, but I've got an odd networking problem - firefox is unable to load URLs. Every URL I try returns a server not found error.
When I drop into bash I'm able to do the following Code: ping 66.102.9.103 ping google.com
However when I try Code: wget http://google.com
I just get a message that tells me that wget is "unable to resolve host address google.com'". This is odd - ping is able to resolve google.com, but wget isn't. I assume that firefox and Konqueror both have the same problem. Could it be cause I've specified the http protocol?
I've added my public key to the remote machine's authorized_keys file, and I can ssh over without password. But when I try to mount the remote share using sshfs it -always- asks for my user's password. I have set sshd_config|PasswordAuthentication no ... and when I mount the share as root it says, "read: Connection reset by peer". My mount is being done as user, so it shouldn't be a root authentication problem: sshfs#bill@droog://media/droogfuseuser,noauto,gid=6,umask=007,cache=no,ServerAliveInterval=15,reconnect,allow_other,comment=sshfs 0 0 I can't mount as user because /dev/fuse is not suid, and I'd rather not set it such.
I'm trying to get BIND running on one of our Fedora 10 servers, since we've had some issues with the DNS from our service provider. This BIND server is not authoritative for any domain, it's merely for caching purposes. I used the sample DNS file at [URL].., and modified it slightly with our internal IP addresses.
I'm getting the following errors when I do service named start:
Error in named configuration: zone localhost/IN: loading from master file master.localhost failed: file not found _default/localhost/in: file not found zone 0.0.127.in-addr.arpa/IN: loading from master file localhost.rev failed: file not found _default/0.0.127.in-addr.arpa/in: file not found
I also tried changing the line directory "/var/named"; To the /var/named/chroot/var/named directory with the same result.
I need to setup 2 BIND servers where I would need to configure automatic fail over where if one of the DNS server goes down other one should be able to pick it up.I was thinking to deploy VRRP solution with keepalived or vrrpd
I'm setting up an ftp server with lucid server. A lot of the folders that should be accessible via the ftp are in different directories (and can't be moved without a LOT of hazzle) and I have to either symlink or mount bind them to the ftp chroot dir. Now I'm wondering which one is the saver variant? My guess is mount bind, but I'm not that familiar with the internal workings of linux and vsftpd (plus for symlinks I wouldn't have to change/create any scripts, just create them once...),
I have 2 servers operating on my home network. One server hosts Bind9 and a few web sites. The other is an Ubuntu repo mirror that I recently configured (10.1.10.26). Inside my network, requests to the olyubuntu.nwlinux domain function as it should. However, external requests do not get forwarded. Instead, they end up at my DNS box at 10.1.10.25. Again, internal requests are forwarded correctly.Any red flags pop-up as to why this is occurring? Firewalls are not an issue.
I can't get a program (wbar) to run directly from my user account, it fails saying "Image not found -> maybe using a relative path?". But if I run su -c "wbar", it shows up and manages to load the image. I think it has something to do with ImLib2 or whatever loads the image. I checked permissions on libImlib2.so.1 and it's world-readable and executable. Can libImlib2.a be causing this problem, set to 644? What else should I be checking?
Having an odd problem running a mysqldump via crontab. I have the script running on other servers and they work fine, so not sure how to actually troubleshoot, but the script looks like the following;
If I run it as a cronjob as root, it finishes in a second and a 20k file is there. If I run it from the command line as root it does the backup (takes a few minutes) but does complete the backup and can be unzipped and read successfully.
I installed freenx-0.7.3-i486-1alien.tgz on the server (with Slack 12.1), all according to instructions. I installed the no machine nxclient 3.4.0-7 on the client (with Slack 12.2). I ran setup and configured. I used the default no-machine keys, not custom ones. Since authentication failed at the beginning, I enabled DBauthentication, and added my user and password, which seemed to allow authentication to occur.
Here is my node.conf: # node.conf # # This file is provided by FreeNX. It should be placed either into # /etc/nxserver/node.conf (FreeNX style) or /usr/NX/etc/node.conf # (NoMachine NX style).....
I hope I haven't missed this in another forum but was tough to search for.While administering my new centos 5.2 x86 server through SSH I am successfully able to issue