Have a headless server running Lucid 64 bit. Everything is working great except for a couple of Cron jos I'm trying to have run at boot.
My user crontab looks like this: # m h dom mon dow command 01 * * * * /home/ceallred/Scripts/SambaBkup.sh @reboot /usr/bin/SpiderOak --headless &The SambaBkup script works like clockwork...
The SpiderOak job doesn't start. It works if I type the command in manually. syslog only shows:Jul 13 22:45:33 RavenWing cron[1010]: (CRON) INFO (pidfile fd = 3) Jul 13 22:45:33 RavenWing cron[1022]: (CRON) STARTUP (fork ok) Jul 13 22:45:33 RavenWing cron[1022]: (CRON) INFO (Running @reboot jobs) Jul 13 22:45:33 RavenWing CRON[1048]: (ceallred) CMD (/usr/bin/SpiderOak --headless &)Yet ps -A doesn't show a running SpiderOak process. It shows after starting manually.
Cron for root does not run,it's 8.04LTS ubuntu server,however Cron runs for other users,Can anyone help?i've checked many times that i have the right syntax.
Ubuntu server 10.04:I had webmin installed on my server, but I found it was using too many resources for what it offered, so I went ahead and uninstalled it. I'm not sure exactly which version it was, but it was the most recent version a month ago. I ran the uninstaller which ships with webmin.Well, now everyday my root account gets the following email:
From root@DOMAIN.com Mon Oct 18 00:42:01 2010 Return-Path: <root@DOMAIN.com> Date: Mon, 18 Oct 2010 00:42:01 -0400
I am running a headless Ubuntu server accessed through Webmin. The server is running 10.4.2 64 bit version. I have a number of cron jobs including a simple back-up job which is: Code: rsync -av /media/server/ /media/backup/backup/ All of the other jobs run fine but for some reason this job which is scheduled to run each day at midnight does not run. If I SSH into the server and run the job manually it works fine.
I am unable to run my cron tab job without root password.I made a little research on the forums and appearantly it should be changed.Must I change "passwd" part with my root password?
I'm trying to setup a cron with screen. the purpose of running screen is because i want to resume to see the status of the job while it is running. anyone can enlighten me how i should setup my cron. I've tested with
I'm with Ubuntu server 10.10. I created a script and put it in /etc/cron.daily. But looks it doesn't run (it didn't generate any log). The following is all I did. Code: root@chonseng1:/var/log# cat /etc/crontab # /etc/crontab: system-wide crontab # Unlike any other crontab you don't have to run the `crontab' command to install the new version when you edit this file and files in /etc/cron.d. These files also have username fields, that none of the other crontabs do. SHELL=/bin/sh PATH=/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin
# m h dom mon dow user command 17 * * * * root cd / && run-parts --report /etc/cron.hourly 25 6 * * * root test -x /usr/sbin/anacron || ( cd / && run-parts --report /etc/cron.daily ) 47 6 * * 7 root test -x /usr/sbin/anacron || ( cd / && run-parts --report /etc/cron.weekly ) 52 6 1 * * root test -x /usr/sbin/anacron || ( cd / && run-parts --report /etc/cron.monthly ) # root@chonseng1:/var/log# ls -l /etc/cron.daily/ntpdate -rwxr-xr-x 1 root root 52 2011-02-18 13:22 /etc/cron.daily/ntpdate root@chonseng1:/var/log# cat /etc/cron.daily/ntpdate ntpdate ntp.ubuntu.com >>/var/log/ntpdate.log 2>&1
I am using my ubuntu server as my home router.Everything is working as expected with one exception.y DSL modem is a POS and every now and again it looses connection to the router. Sometimes it needs to be reset and sometimes it does not.Either way, when this happens my ubuntu server needs to reacquire an IP from my ISP. If it screws up when I'm at home it's no big deal, but if it happens when I'm not around my housemates have taken to hitting the reset switch on the server. I'm not a big fan of this so I wrote a script to ping my ISP's gateway. If it's unavailable it bounces the eth0 interface and tries to get an IP. I am running this script every couple of minutes in a cron job. Now I'm getting syslog entries like
Code: Aug 9 20:31:01 portal CRON[9602]: (root) CMD (/opt/ChkAndFixNetwork.sh) every few minutes. This is annoying and makes the logs useless for troubleshooting. I
I'm creating a script all worked fine in the command line. But not work in the cron. Below you could see the script
[Code]...
So far I found when I use corn following part not working, nothing goes to the processedfiles file. ls -l /var/lct/mou2/processed | grep $TODAY | awk '{print " " $8}' > /home/trans/mou/processedfiles ls -l /var/lct/mou2/processed | grep $YESTERDAY | awk '{print " " $8}' >> /home/trans/mou/processedfiles This work perfect in command line. Corn job and command line use by the same user.
I've read a lot on how to setup a task in crontab. I think I understand how to edit the file, but when I try to save the changes, it says "no crontab for xxx." Then it says that it cannot create a new crontab. I have ubuntu desktop 9 running as a webserver. I've read the details in these posts and it isn't helping.
I have a cron job to run every 15 minutes. The cron job is running per /var/log/cron. It lists the job every 15 minutes.
[Code]...
I do not see any errors. I also looked in /var/log/messages for any errors and it does not list any. I added a line to the script that is running to send an email to me to see if it is running or not. When I run the script from the command line, the script runs without error and I get an email. I have searched and can not find what it going on. Is there some where else errors might be? Could there be a permission issue? The cron job is not being added by the root user.
Code: #!/bin/sh chown -Rc steven:users "/home/steven" I have more lines in the script but each line does the same error. This is run in a cron job but it always returns.
Code: chown: cannot access `/home/steven ': No such file or directory But it is a valid directory I have even tried adding an / and even a space on the end and still same error.
I recently set up exim so that I could test emails sent from my php sites. I have now been getting daily emails that either report errors from cron jobs or report a failure to send emails to root@cory-laptop or cory@cory-laptop about cron errors.I first of all don't want mail sent to the above addresses, because they are obviously not actual email addresses. Second, I'd really prefer not to get emails when an email failed to send to these addresses.
I am running a series of operations on executable matlab scripts which are connecting python and fortran scripts. I schedule the executables in cron. The problem:
- My original matlab scripts work perfectly. - My executables work perfectly if I run them from the terminal. - Using the gui scheduled tasks and running the scripts once from a button, everything also works fine. - But when I leave the scheduled tasks run on their own, I get an error! The error can be that the script hangs in general (I have some text logs exported every step to track the progress), or I get an error which never appears when I run the script with any of the other mentioned ways. - I tried both cron command prompt or the GUI scheduled tasks - I am running on Ubuntu 64 bit
I'm getting the following messages sent to my root account mailbox. It appears to be reporting an issue finding ntpdate, however when running the following command:
aptitude show ntpdate | grep State
It shows as installed and any ntpdate commands work with no issues, so I have no idea why this is getting regularly reported?
From root@mydomain.co.uk Wed Feb 03 18:20:01 2010 Return-path: <root@mydomain.co.uk> Envelope-to: root@mydomain.co.uk Delivery-date: Wed, 03 Feb 2010 18:20:01 +0000
I wrote/cobbled-together this nifty sendmail script to read some logs and take some disk stats. basically i'm reporting on rsnapshot. When I run it as
Code: sudo /etc/rsnapshot/mailSta.sh Everything works wonderful and the emails arrive as expected and fires off an email to two accounts at a remote server
I made a Bash script that is fired by a Cron job every morning. It dumps an SVN backup on some Samba shared drive. I would like to know how I can make sure the job worked correctly without having to verify the shared drive every morning. Right now, I take the job's output, save it to a log file and send this file by email. But the ouput isn't so great.
I'm running a cron job every night to dump a MySQL database to an external hard drive. It works, however when I check on it the following morning the external is no longer mounted and the XFS log file is corrupted. If I run
Code: xfs_repair -L /dev/sdf1 It works, but then I get these issues: Code: XFS: Filesystem sdf1 has duplicate UUID - can't mount I can reset the UUID, but it's difficult to have to do this every day.
I'm running cygwin 1.7 on a vanilla win2k3 r2 server. I've put together a simple crontab that I'll paste in her in a sec. The crontab entries call several shell scripts that do either scp's or rsync's from linux servers to this win2k3 box under cygwin.
The problem is that the scripts work and run fine from the commandline. Just not when called from the crontab. I'm running from the root equivalent administrator UID. The crontab log files are empty of errors.
crontab is:
Like I said, the scripts execute fine from the command line.
I run a Fedora 9 server at home, to host an "old school" MOO.To back up the database, I scheduled a cron job - and got some help with the script.I don't fiddle with stuff on the server much, because I don't really have a clue - leave well enough alone.But now I'm without backups. (I'd prefer to get this working, as it seemed pretty simple, and worked well for so long).
It is little weird now that cron job is not working as i have set. I have set weekly job to send email but i am receiving multiple emails daily. Below is my code; # sends email every monday at 4:00 am 00 04 * * 1 /usr/bin/ruby /home/mbm/www/current/script/runner /home/mbm/www/current/app/models/add_to_delayed_job.rb -e production
00 -> Minutes 04 -> Hours (0-24) 1 -> Days / 0-6 / Sun - Sat
Sending email is fine but the time is not working so far as set.