I have a XEN DomU VM running Lenny and set up using debootstrap. I also have the TimeWentBackwards issue sorted using: [URL]. The problem is that my cron jobs just don't run. I've set up a cron job that simply echoes a date string to a file as a tester but it's not running. I've tried it under the root crontab, in /etc/crontab and set up in /etc/cron.daily. Is there something I'm missing?
I have added some executable scripts to /etc/cron.daily but don't get the stdout/stderr output from them as mail (or anywhere else I have found). At least one of them is running (because I can see that it has added a file to the disk).
The peculiar thing is that I do get the output from /etc/cron.daily/0logwatch (part of the logwatch package) as an email each day.
The MAILTO line in /etc/crontab is "MAILTO=root" (unchanged from default). Same for /etc/anacrontab.
I do have an alias at the end of /etc/aliases which redirects root's mail to my own account, but this alias works fine for mail I send manually. (It also appears to work fine for the output from the file /etc/cron.daily/0logwatch.)
I have this in my crontab (along with another line before it):
Code: * * * * * /bin/echo "Does this work?" > ~/test ...and the file doesn't show up. I have a blank line after that one in the crontab. I have no /etc/cron.allow or /etc/cron.deny files. I'm running Ubuntu 11.04. /var/log/syslog shows "BEGIN EDIT" and "END EDIT" etc. when I run crontab -e, but doesn't say anything about trying to run commands. Anything I might be missing?
I want to add two separate tasks/jobs to Cron that needs to run every 10 minutes, but I would like the second job to kick off only 5 minutes after the first job. Is there any way this could be done within Cron?
run a cron job to download my email logs to my laptop. But the question is HOW? Im not sure how to write the php script for the cron job. the file is on "computer A" and the file is setup to chmod I just need to know how to write the php to "access" the other computer and then download the file to a certain file on my laptop.
Have a headless server running Lucid 64 bit. Everything is working great except for a couple of Cron jos I'm trying to have run at boot.
My user crontab looks like this: # m h dom mon dow command 01 * * * * /home/ceallred/Scripts/SambaBkup.sh @reboot /usr/bin/SpiderOak --headless &The SambaBkup script works like clockwork...
The SpiderOak job doesn't start. It works if I type the command in manually. syslog only shows:Jul 13 22:45:33 RavenWing cron: (CRON) INFO (pidfile fd = 3) Jul 13 22:45:33 RavenWing cron: (CRON) STARTUP (fork ok) Jul 13 22:45:33 RavenWing cron: (CRON) INFO (Running @reboot jobs) Jul 13 22:45:33 RavenWing CRON: (ceallred) CMD (/usr/bin/SpiderOak --headless &)Yet ps -A doesn't show a running SpiderOak process. It shows after starting manually.
I recently set up exim so that I could test emails sent from my php sites. I have now been getting daily emails that either report errors from cron jobs or report a failure to send emails to root@cory-laptop or cory@cory-laptop about cron errors.I first of all don't want mail sent to the above addresses, because they are obviously not actual email addresses. Second, I'd really prefer not to get emails when an email failed to send to these addresses.
I have a file server that is an NIS client. User home directories are auto-mounted from another server. Users are not able to run cron jobs. The /var/log/cron logs says: Jan 6 14:33:01 inclination crond: (username) ORPHAN (no passwd entry)
I tried adding an entry in /etc/passwd, /etc/group, and /etc/shadow but the test script I have still doesn't seem to run although cron thinks it did. The log entry for these attempts looks like: Jan 6 15:30:01 inclination crond: (username) CMD (/home/username/test-script)
Now root does seem to be able to run cron jobs with no problem. My test script works fine from cron when runs as root. I suspect there's an issue with how I have NIS configured but I can't find any info. The test script runs fine from the command line when logged in as myself.
By default, cron jobs are scheduled to run at 4:40 IIRC. I have my computer set to suspend after half an hour of inactivity so, except for some all-nighters, pretty much every day at that time my PC is suspended. What happens to the cron jobs? I don't suppose they run while the suspension is in effect. Do they get run when the computer wakes up or does cron wait until the next day at 4:40 to run the jobs?
Just curious if this is possible. What I want to do is setup a rsync job to backup my laptop to my personal file server(same LAN), but I want it do do this at 3AM while I'm sleeping. However I always close my laptop lid when I'm done using it which puts the laptop into suspend mode since that's how I configured my power options in gnome. Since I don't want to leave my laptop powered up all the time, I was wondering if it is possible to have a cron job scheduled that will wake the laptop up(out of suspend mode) and run my script/backup job, all without opening the lid of the laptop, and then put it back into suspend mode when it's done. Is this possible, and is it as easy as scheduling a cron job or is there some other scripting/configuration/trickery that I need to do to accomplish this?
Also, my laptop's BIOS has the option to power on at a scheduled time if needed, but I'm not sure if that would work with it being suspended(hibernation is not an option since my entire HD is encrypted with LUKS and I would have to be present to enter the password to boot the system.).
My server... its CentOS (redhat alike). I need to make the server disable apache server then run mysql command to optimize the databases then re-enable apache server in a specific time daily. Is it possible?
I have a Cent OS dedicated server, not sure what version though as I'm new to Linux. How do I find out what version I have? Is there an anti virus or security package that I can install on my server which can use Cron Jobs to do a scan every 12 hours.
I am having a shell script which runs perl jobs.The script is starting the perl jobs when it is executed manually from the command line , but when the same script runs from crontab it is not starting the perl jobs.I have these things in the begining of the script
What you do if the job takes a long time to finish and you don't want to wait.Say, I ssh to a remote server from my laptop and start a long-running job. Then few hours later I ssh again and inspect how did job run, its uotput and etc.