Having an odd problem running a mysqldump via crontab. I have the script running on other servers and they work fine, so not sure how to actually troubleshoot, but the script looks like the following;
If I run it as a cronjob as root, it finishes in a second and a 20k file is there. If I run it from the command line as root it does the backup (takes a few minutes) but does complete the backup and can be unzipped and read successfully.
I've successfully mounted a network share with mount.cifs for the past 2 years using fstab with credfile.
[Code]....
Yesterday I moved this system to a new datacenter, but did not alter fstab or the credfile. The //server/share directory has IP rules in place, but this was updated with the new system IP while we moved the system. Now, I am mysteriously unable to automount //server/share. The local error is 13 (permission denied). The Windows server we are mounting returned a code that is defined as "username is valid but password is incorrect" Again - no changes (content or permissions) were made to my credfile or fstab entry. I've restarted netfs a few times, including rebooting the system twice. What is baffling is I can successfully mount //server/share via command line: Code: mount -t cifs //server/share /mnt/mycooldir -o username=foobar,password=1234
The username and passwords are identical in credfile and the mount options - I copied & pasted username / password from the credfile itself.
I have two Roboards that need to communicate with one another. I have sucessfully installed Ubuntu 9.04 using the 386 kernel based on instructions found online.I can use the GUI to set up an Ad-Hoc network using one Roboard ("Create New Wireless Network"), call it Manet, and connect to "Manet" using the GUI on the other Roboard and ping/ssh between the boards. However, when I try to create the same wireless network using the command line using:
ifconfig wlan2 down iwconfig wlan2 key off essid 'Manet' channel 5 ap any ifconfig wlan2 up
I don't see any entry in iwconfig under "Cell". It keeps giving me "Not Associated"When I try to enter the same information on the other Roboard to try to connect to the network, it doesn't connect and under iwconfig, I don't even see any entry for ESSID. Could somebody please tell me why it works when I use the GUI and not the command line.
Does anyone know why Nautilus would allow me to login to a server but not display any files, but when I log in via the 'ftp' command-line binary, it works and allows me to display all the files, make directories, etc.? I was able to login via FileZilla too.
Here's part of the log: Code: Response:220 Microsoft FTP Service Command:USER ---- Response:331 Password required for ----. Command:PASS *********** Response:230-Welcome to the Alentus FTP server. Response:230 User ---- logged in. Command:SYST Response:215 Windows_NT Command:FEAT Response:211-FEAT Response:SIZE Response:MDTM Response:211 END Status:Connected Status:Retrieving directory listing...
I'm trying to configure sending mail from my site under Ubuntu 10.04 LTS.I have installed LAMP and postfix. To test mail I created the file test.php with code:
PHP Code: $to = 'mymail@gmail.com'; $subject = 'The subject';
How can I verify JavaScript on a site (not on the same server) is working from command line access of the URL?
On checking some of what was written in 2005 I find most of it isn't still relevant (if it was then - I noted the w3m-js stuff mentioned in 2005 hadn't been updated since 2003).
Since then a couple of times I've looked into it but didn't make much headway but also didn't spend a lot of time on it. So far as I could tell from other posts I'd seen others hadn't either.
Recently I found out about Mozilla's SpiderMoneky js library and that it could be linked into elinks text browser and did same on my RHEL5 system.
I found Spidermonkey rpms in the EPEL (Fedora sponsored) repository: js-1.70-8.el5.i386.rpm js-1.70-8.el5.x86_64.rpm js-devel-1.70-8.el5.i386.rpm js-devel-1.70-8.el5.x86_64.rpm
I was able to download and install using rpm then compile elinks from source to include the above. Running elinks version after that shows the ECMAScript support is compiled into the new elinks binary. Also setting various ecmas* flags in the elinks.conf does have effect so clearly it is using them. (As opposed to another system where I use the original RHEL5 provided elinks RPM it complains about the ecmas* stuff being invalid options.)
It appears that by default it has this enabled but just for good measure I added:
set ecmascript.enable = 1.
Despite that when I pull up the page I see:
This product requires use of a browser that supports JavaScript
As a test I set the above to 0 on the off chance it was backwards and saw the same thing.
I then set the following:
set ecmascript.error_reporting = 1
On pulling up the page I now see a popup box:
Quote:
JavaScript Error:
A script embedded in the current document raised the
following exception:
TypeError: Window.Focus is not a function
OK
That seems to suggest that despite the earlier message about needing a browser that supports JavaScript that it is in fact executing at least one JS but that it is getting an error. Can anyone confirm that?
Does anyone have any idea what I�d need to do to deal with the Window.Focus message?
Essentially the site is a login page and even though I am able to input username and password after opening in browser when I tell it to POST it simply returns to the same login/password page.
And of course it doesn't have to be elinks/spidermonkey. I'd be interested in finding out if anyone knows a way I can verify a site that is running JavaScript is actually responding properly from command line rather than via a browser?
The SpiderMonkey stuff from Mozilla when compiled into elinks is supposed to do that but given my results Im not sure if it is or not.
Most of the documentation I can find on JavaScript stuff appears to be aimed at developers testing their own JS code from command line rather than accessing a web page with JavaScript.
I'm writing a Bash script to take IPTC keywords from a text file and write them, via Exiv2, to several (first batch is 100) JPEG files in a single directory. The script has one while loop inside another while loop, both terminated, but I'm pretty sure that's not my problem. I think it's how I'm incrementing the "counter" variable, although it could also be the method of parsing the text lines from the file (using cut with delimiters that have worked fine in simpler scripts).
Here's the code as I've worked it up to this point.
Code:
And yes, "keywords" checks out in Crimson Editor, Emacs GUI and nano as an ASCII file with UNIX line endings. No issues on that score.
Feeding each line consecutively into a terminal (excepting the exiv2 command) works fine: each variable echoes with the part of the text line used as a variable value as it should, even when the b variable is incremented the quick&dirty way (up arrow three commands and hit enter).
Running the above script in eval mode (sh -x) stalls after setting the b variable to one and reading in the first line of text. I'd like to know why. I'd also like some advice on another reliable method of parsing the read-in lines.
So, in finishing my nFlux slack current edition.I have set it up for users to do certain things in console and one of the things I want is a way to view slackbook-2.0 in runlevel 3 console.I cant find a pdf reader that works in command line mode and I cant figure out how to either convert slackbook 2.0 pdf into html/text Or find a slackbook download that is html or text?I tried converting it using pdftotext, which didnt work very well So, I need a command line pdf viewer or a converter that works good?
Writing script to create backup of file by adding datetime to file name. Basically test for file presence if there, cp with datetime then rm original cp works fine from command line but get cannot stat `full path to file': No such file or directory
Code:
Here are the errors: cp: cannot stat `~/html/CVP_dadamail/.dada_files/.logs/errors.txt': No such file or directory rm: cannot remove `...': No such file or directory
The for statement is a placeholder as I have same file to backup out of several directories. using "bash -x scriptname" -OR- inserting echos, I can see I've constructed the strings properly. Believing it might be related to the hidden directories, I tried setting the shopt "glob" options to no avail.
Ultimately I'll add the other directories to the for loop and then run this from a cron job, so if you see potential pitfalls knowing I'm headed in that direction...believe construct would be
I have a question about using crontab with /etc/crontab...
I had a cron job that I needed to run as root. At the time I thought that sticking it in /etc/crontab would be a good idea. However, I used the crontab command to edit /etc/crontab, which I guess is not standard procedure? Specifically, I configured /etc/crontab as my local user's crontab (i.e. sudo crontab /etc/crontab) then added my cron job as I would a local user crontab (i.e. sudo crontab -e).
Originally, my cron job looked like this:
30 * * * * root /my/batch/script &> /dev/null
After adding the new cron job I started seeing errors. Something to the effect of "can't find command root" or something similar. So I removed the 'root' user definition from the cron job and the job started running fine. However, because this is /etc/crontab, there are other system related cron jobs that have been defined to run under the root account (e.g. "17 * * * * root cd / && run-parts --report /etc/cron.hourly" runs as root, etc.). So these pre-existing system cron jobs, which up until now have been running smoothly, are now generating "can't find command root" errors. But I think that the system cron jobs _are_ successfully being run someplace because logrotate seems to be working.
So what I _think_ is happening is that /etc/crontab is being run twice: once as the system crontab, and once as my sudoed local user's crontab. When I run crontab -l I see nothing, but when I run sudo crontab -l I can see the contents of /etc/crontab. I am reluctant to delete my sudoed local user's crontab, because then in the process I would be deleting the system crontab, and I do not know how I should restore the system crontab's contents. (I am still not sure as to the most appropriate way to edit the system crontab).
How can I get out of this mess? I want /etc/crontab to go back to the way it was before--running _once_ as the system crontab. As for my new cron job, I'm willing to reconfigure it anywhere so long as I am still able to run it as root. Any ideas? (I am using Ubuntu 8.04 Server LTE)
I want to write the date & time and a text string to a file from crontab.The following line works fine in the CL:echo $(/bin/date +"%F %T")" Some text" >> /home/me/foo.txtI installed in crontab and no text appears in the file that it is redirected to.The crontab entry looks like:* * * * * echo $(/bin/date +"%F %T")" Some text" >> /home/me/foo.txtTried a version to just write to stdout....* * * * * echo $(/bin/date +"%F %T")" Some text"No date, time or text appears at the command line
I use crontab to execute a bash script every minute and the output is redirected to a log file.In that bash script there is a screen command to start a screen session.When I execute the script manually it works perfect, but with crontab, the screen command is not executed, but the echo's are. Because they are written in the log file...Here's my script:Script removed for violation of LQ Ruleso the script runs perfect when I do it manually, but not with cron...Here's my crontab (sudo crontab -e)
# m h dom mon dow command */1 * * * * sudo /home/laurent/Games/Cod4/StartCod4.sh | egrep 'running|started' >> /home/laurent/Games/Cod4/Serverlog.log
I am using Linux 64 bit Redhat Linux. I am trying to setup simple crontab as follow...1. Edited crontab file using crontab -e2. Listed the file once to verify it using crontab -l. This will display as.. 18 5 * * 2-3 ksh $HOME/testScript.sh > $HOME/testscript.out3. Logged in a root and restarted cron deamon using "/etc/init.d/crond restart"As per my understanding now my testScript should start running at 5:18 am Thuesday
i've gotten my fedora 12 to the point where i can run python3 scripts from command line and can call up python 2.6.2 idle with the command 'idle' from command line. what command will call up python3 (3.1.2 to be exact) idle?
As root, I use crontab to run mirrordir to backup directories. Everything gets copied over properly, but owner information isn't preserved and root is the owner of all the backed up files. I can deal with that, but crontab reports tons and tons of chown/chgrp errors for mirrordir every time I do back ups--which is every day--and the multiple emails to root of thousands of chown/chgrp errors is very annoying. The error is "Operation not permitted," but that doesn't make sense to me because the job runs as root (right?) and clearly the job is permitted to create the backup files, so why would it fail to chown and chgrp?
I've had the exact same setup on another server for years, and crontab has always run mirrordir without error. Any suggestions how to clear the errors on my new server?
I have written a simple backup script, and added it to CronTab, but it doesnt execute at all. Here is my script: [URL]...And my CronTab entry: 0 */2 * * * root /home/server/Scripts/backup.sh
I was having so much trouble with ubuntu 8.04 that I deleted off my computer and did a fresh install of 9.10. I downloaded the 64 bit iso from the internet, burned it to a disc and installed it. When it asked if it should be the server version I thought I said no. Now when I boot it only boots to the server version and all I get is command line. Can I get out of this and get my regular ubuntu screen back or did I install the wrong version?
Is been a while since I was last here requesting help. Now I need some of that LQ magic. I have a script file that extracts data for a date range, create a zip file for the data and then ftp it to a remote server. Now the script is working perfectly if it is run manually, but now I want it to execute automatically. So I use the crontab, but for some reason the script fails when run from the crontab.code...
Whilst I have used GUI-based Linux distros for the last few years, I am now struggling somewhat when it comes to setting up a server from the command line. All the howto's I've read so far tell you how to set up certain things, but I don't know what I actually need to set up.Rather than asking the same questions all over again on a forum like this, and generally being a bit of a noob, I wonder if there's a resource somewhere that someone could direct me to in order to know what I need to do to set up a fileserver.
The hardware is all done, and I will probably set up a router distro like IpCop to manage the network, but I'm rather lost as to what packages I need, and how to know what the server is/isn't doing.I've used plenty of command lines in the past (ms dos, amiga dos, BASIC programming etc), just not a linux command line. All I've ever done is mount a few drives, use nmap and started x
I was using Red Hat 7.3 forever and decided it was time for a change. I went to Fedora 10 but it was really buggy. CentOS 5.2 is VERY stable.Here is my problem.The server is command line only -- I tend to hate GUI"S.I setup Samba no problem disabling the ports needed through the firewall and that was straightforward.CUPS is a nightmare for me since CentOS locks down the cupsd.conf and then the firewall does its thing. I allowed port 631 through the firewall but then got lost on the cupsd.conf. It's been too long and the old redhat one won't work with CentOS (not surprising since it's a VERY old system) straight-forward CUPS tutorial for a command line interface. I just need it to be:
I installed exim4 package on my debian server as MTA. Now I want to send mail with this by SMTP. But when I connect to this via telnet, it returns this banner:
Whenever I try and use php from the command line, I get this message:PHP Warning: PHP Startup: Unable to load dynamic library '/usr/lib64/php/modules/imagick.so' - /usr/lib64/php/modules/imagick.so: cannot open shared object file: No such file or directory in Unknown on line 0The problem isn't the missing module, it is the warming message itself -- it is breaking scripts that rely on the output.I tried shutting off errors, changed my error reporting to:
I just got done installing Ubuntu Server 10.04 on a home server. Now what should I do? I was hoping to have 3 computers connected to it so that we can all share files wirelessly if that's possible.Does Ubuntu Server have a GUI or command line only?
I am accessing a linux server remotely from my putty. I started the server and now I want to close the command line. when I do cntr+c or cntr+z it kills my server aswell. how will I close my terminal without closing my server? I tried cntr +d but it is not doing anything