I have prepared a script which will login to each server and search for a keyword.I want output on same machine from where m running script.When i try to run command on any machine.. It works well.
I have been looking for a while now, but I keep getting 403's for maps but not for files... So if you go to http://gmod.ws/ you get the error but if you go to http://gmod.ws/index.php you don't.I don't see where the problem is.We're running a CentOS 5.5 box.
my web host runs a linux server, and when i try to load a file in my browser (which i have uploaded in my web space) with non latin words it gives a 404 error (file not found). for example i have uploaded mydomain.com/νεο.html the word "νεο" is non latin. so when i try to reach this document from my browser i get the error.
I am using SUSE 11.0 KDE 4.0 . My system has root user, one normal user and both are me only. I am using only 'user account' only.I used to install softwares from 'user account' by using YAST.
Now my problem is - when i am trying install a rmp file through YAST, it ask root password by a dailog window. when i entered the root password it is giving error
Quote:
The program 'SU' not found. </br>Make sure your path
I had attched a screen captured image of that window.
And then after i had to be ignore that message. Installation just starts and then gives error, installation fails. so now i am not able to install any softwares through YAST !
I am having a 1 gb of mysqldump. When i try to restore this dump file, it is giving error " ERROR 1153 (08S01) at line 289: Got a packet bigger than 'max_allowed_packet' bytes". And with this i am not able to restore the database. mysqld version is 5.0.77.
I'm trying to setup a small network between my old and new laptops to transfer my personal data. They are now linked with a crossover cable and they see each other.The old one has a dual-boot setup with WinXp and Ubuntu 9.10.The new one with Win7 and Ubuntu 9.10.I tried samba but it was very slow even using Windows in both computers: maximum transfer rates were about 1,5 Mib/sec.I tried SSH using ubuntu on both pcs and it is reliable and much faster, 5 Mib/sec. But I wanted more...I installed the NFS server on the old one and exported the NTFS partition where my data resides with sync and ro options.
I installed the NFS client on the new one and i'm able to mount the remote partition.Now, when I transfer my files I get very high speed, more than 10 Mib/sec but after a while I get a "Stale NFS file handler" error but I really didn't touch any file in the old pc and the connection is always up.Searching on the web I found that NFS had some troubles exporting NTFS partitions in the past but should be fully compatible with them since the last versions of ubuntu.
I am using openSUSE 10.3.When I install software from tarball then to record time required I send output of date to beg.txt(when installation begins) and end.txt (when installation finishes).How can I append output of date to a file so I don't need two files?
I have a question about using crontab with /etc/crontab...
I had a cron job that I needed to run as root. At the time I thought that sticking it in /etc/crontab would be a good idea. However, I used the crontab command to edit /etc/crontab, which I guess is not standard procedure? Specifically, I configured /etc/crontab as my local user's crontab (i.e. sudo crontab /etc/crontab) then added my cron job as I would a local user crontab (i.e. sudo crontab -e).
Originally, my cron job looked like this:
30 * * * * root /my/batch/script &> /dev/null
After adding the new cron job I started seeing errors. Something to the effect of "can't find command root" or something similar. So I removed the 'root' user definition from the cron job and the job started running fine. However, because this is /etc/crontab, there are other system related cron jobs that have been defined to run under the root account (e.g. "17 * * * * root cd / && run-parts --report /etc/cron.hourly" runs as root, etc.). So these pre-existing system cron jobs, which up until now have been running smoothly, are now generating "can't find command root" errors. But I think that the system cron jobs _are_ successfully being run someplace because logrotate seems to be working.
So what I _think_ is happening is that /etc/crontab is being run twice: once as the system crontab, and once as my sudoed local user's crontab. When I run crontab -l I see nothing, but when I run sudo crontab -l I can see the contents of /etc/crontab. I am reluctant to delete my sudoed local user's crontab, because then in the process I would be deleting the system crontab, and I do not know how I should restore the system crontab's contents. (I am still not sure as to the most appropriate way to edit the system crontab).
How can I get out of this mess? I want /etc/crontab to go back to the way it was before--running _once_ as the system crontab. As for my new cron job, I'm willing to reconfigure it anywhere so long as I am still able to run it as root. Any ideas? (I am using Ubuntu 8.04 Server LTE)
I'm using postfix to send some mail from my cloud server via a web page on my website that uses PHPmailer to do SMTP authentication. My cloud server has a dedicated IP for all inbound and outbound data. The problem I am having is that it sends roughly 75 - 100 messages successfully before it returns the following error for the remaining few hundred recipients in my list:
SMTP Error: The following recipients failed: example@example.com Mailer Error [URL] SMTP Error: The following recipients failed: example@example.com SMTP server error: 5.5.0 Sender already specified
Through shell script iam sending mail, i want to send mail to multiple recipient,can any one help me how to send mail to multiple recipient.Iam sending mail through the below given in shell script to one recipient.
Every time I boot up ubuntu I usually open 3 terminal windows and ssh into the same server. I would like to either click a shortcut, or run a single terminal command that will do the equivalent.
I came across the "gnome-terminal" command, but I was unable to get it to trigger an ssh command.
Ideally I would like to have a script that I pass in the number of windows I want to open and the server I would like to ssh into for each window.
I have a pretty fresh install on Ubuntu 10.10 Server, with the LAMP bundle installed. Anyway, I just restarted the computer after changing the /etc/network/interfaces script to connect via dhcp. Now when I boot into the server, there is no command prompt and seems to hang there forever, with one blinking cursor in the top left hand corner of the screen. It will not respond to any input besides ctrl-alt-delete. I was able to press control-alt-delete and the computer automatically restarted itself, but it returns to the same state after each reboot. I booted from the grub menu into recovery mode and made sure the /etc/network/interfaces script was correct, and it is. Also I don't think an error in this file should cause the computer to not boot into the command prompt.
When ever i am giving the command service nfs restart in fedora14 iam getting the following:
And when typing the command service portmap restart in fedora 14 i am getting the following:
When i tried this in redhat and centos it was working in my class room but when i am trying it in real time it is not working so what does it mean and how to resolve the issue.
I've been dual-booting win7 and ubuntu on my comp for a while, but didn't bother updating grub. I just updated to grub 2, but it wouldn't let me boot and said my hard drive was missing, also giving me a grub safety command line.
I have changed nothing about the configuration of my checkgmail except for the icons. And yet, whenever I click on the subject of an email, it gives me a 404 error. When I use the links (delete, spam, archive, mark read) it does what it is supposed to do, but it does not allow me to view the content of a message. Does anybody have any idea of what could be causing this strange behavior? Gnome. 10.04. x64.
In oreder to run an application software on RHEL 3 ES server, I created a link forcefully using following command from root id: # cd /lib64/tls/ # ln -sf libc-2.3.4.so libc.so.6 before that I copied file libc-2.3.4.so from a workstation with OS RHEL 4 WS so that a link can be created. Now I am unable to run any command except cd & pwd and it gives error messaage as given below: ls:relocation error:/lib64/tls/libc.so.6:symbol _rtld_global_ro,version GLIBC_PRIVATE not defined in file ld-linux-x86-64.so.2with link time reference.
Before running this command libc.so.6 was pointing to libc-2.3.so file in path /lib64/libc-2.3.2.so. I am now unable even to open a new window on the server.Please send me some solution as early as possible because this server is running production data and many users are runnig application on this server.
I've got a DNS problem I cannot figure out. I've been digging and pinging all day and I've waited for caches to expire and I'm not having much luck.From my desktop mac, I can't seem to resolve mail.port-o-call.net:
I'm running a hybrid graphics card (AMD Radeon HD 6630m) so naturally I'd like to have full functionality between the two.
Here's what I've done so far: Downloaded the .run file from AMD's website Built packages for both AMD and Intel side of the card Unpackaged said packages
I used ATI Catalyst's GUI to install (in hindsight terminal might have been better and more informative) and it is now saying
After installation of the generated package, it is important to save your X Window configuration file, and then run 'aticonfig --initial' (in a console window) to complete the basic configuration, please run aticonfig from a console window or AMD CCC:LE from the Desktop Manager Menu
So how and why do i save my X Window config and when I give the aticonfig command it says no adapters detected
I am trying to learn it, understand it, and maybe even become proficient with it (one day... ). After trying a few different flavors, I settled on Fedora XFCE, and installed v14 (i686) in a dual boot setting with Windows. I am running it on an HP laptop.I have had an issue ever since I installed a rather large list of updates a few days ago. One of which was a Linux kernel update to 2.6.35.6-45.fc14.i686 (not really sure if that is significant, but I noticed the issue after selecting this kernel version in the boot menu). Now, anytime I try to add software, or run the update utility I get the error message:Quote:You have failed to provide correct authentication. Please check any passwords or account settings.
More Details:Failed to obtain authentication.There was no dialog to provide my password. I do understand there are alternatives to installing and updating software using yum, and I have been doing that a bit, however I would really like to understand why I am receiving this error. I have seen a few related posts as well as a some similar issues in a bug report, and testing session:bugzilla.redhat.com - 638344General a similar gpk-application issuecannot use Add/Remove software through VNCMy questions are:Is this a bug with one of the updates I downloaded?Is there a way to revert the changes from that update, or do I need to reinstall Fedora?Is this an issue with the User account I created? Is there a general set of options I should enable for myself as the primary user?
I have a tar file but when I try to install it is giving error. Here's what I did. Code: tar xjvf live.tar.bz2 cd live ./configure bash: ./configure: No such file or directory I have build-essential installed
I have a Ubuntu 10.04 desktop VM that is running torrent software, and it is saving the files to a samba share (which is actually the host it is running on). I have also tested from another Ubuntu box on the network and it experiences the same problem.
The problem I am having is that any torrent client on a Ubuntu machine or VM can't save to the samba share, but a WinXP torrent client can download the same torrent and save to the share with no problem.
I created a public share, and the problem seems to be that I cannot edit files after I create them. It seems that samba is ignoring the create mask and creating files as rw-r--r-- instead of rw-rw-rw-
I even tried adding the "unix extensions = yes" option and that didn't seem to work either.
Code: [global] # map any unknown username to nobody so login succeeds, don't forget to run: # smbpasswd -an nobody # so that it will work!
I have dhcp3-server (isc-dhcp-server) installed on my Debian and now I got a question about how it's giving the IP addresses to new devices.
For example: I connected my laptop and dhcp server gave me 192.168.1.5 address. Will it always give me the same ip address when I connect my laptop or it will eventually change after some time (week-month)? If it's not changing it, then I am wrong about this.
BUT...If I am somehow correct and if it will change in a week and give me another random (like 192.168.1.8) IP even I won't change my laptop network adapter, is it possible to configure dhcp server to always give same IP address depending on what MAC it is?
To make it clear, I want that when I connect new device (new laptop/pc) dhcp server would give it random IP but same time it would note the MAC address and never change the IP on that MAC.
I know about MAC filters, but setting filters is when you know MAC address since beginning and want to assign IP for it, but in my situation I don't know the MAC address.