A couple days ago I noticed cron stopped working and now I can't create a new file using crontab -e. When I hit crontab -e I get the editor but after saving nothing comes up under crontab -l and the file is blank when I reopen it. I don't have a cron.allow or a cron.deny.
I've tried to create a crontab that renames a fil extension every minute. This is what I've tried:*/1 * * * * / rename /home/bodil/Avräkningsfiler/'s/.txt/dat/' *.txtBut it doesn't seem to work.I've also tried to create a .sh file in the catalog itself with the command executing directly in the catalog but even though running the command by typing it in, it doesn't run automatically by cron
I am using Linux 64 bit Redhat Linux. I am trying to setup simple crontab as follow...1. Edited crontab file using crontab -e2. Listed the file once to verify it using crontab -l. This will display as.. 18 5 * * 2-3 ksh $HOME/testScript.sh > $HOME/testscript.out3. Logged in a root and restarted cron deamon using "/etc/init.d/crond restart"As per my understanding now my testScript should start running at 5:18 am Thuesday
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh
i know the subject is hard to understand but i did my best with it. the problem: i have 2 HDs on my Ubuntu OS, the first is 1TB and the other is 500GB, i want to do a backup with rsync from my 1TB HDs (in with the Ubuntu is on it) to the 500GB.
on the 500GB i have a partition called Backup. what i did: i have created a user crontab that dose two things: 1. mount the Backup partition from the 500GB to a folder called /~/Backup on my home directory which is on the 1TB HD.
2. i wrote the proper rsync command for the backup to go.
i want to create a crontab job by bash scripts:test1 cat /opt/shell/test1crontab on.but i do not want to use this command line ( crontab -e )to add this job.
is it possible disabling a crontab job without deleting the crontab description entry (by crontab -e)?I could also accept to change the entry itself. Now it's:0 0 * * 0-6 /home/me/cron/script.csh
I am trying to run rsnapshot from cron via root's crontab file (crontab -e). If I run rsnapshot from the command line with sudo it works perfectly, however, if I run it from cron:
Code: * * * * * /usr/bin/rsnapshot hourly >/tmp/crontab.out 2>/tmp/crontab.err This does not work. The crontab.err file shows:
i want to do scheduling using /etc/crontab file instead of using crontab -e that is crontab command on the terminal.i am appending to the crontab file in the /etc directory but the scheduling is not happening
I attempted to run various cron jobs as root - just general server make-tidy stuff. But if I edit the /etc/crontab file directly, these tasks don't get run. However, if I use crontab -e as root, these jobs appear to work. Any idea why this might be the case? Also, I understand that user cron jobs are stored in /var/spool/cron/<user> (including root). If this is the case, what's the purpose of the /etc/crontab file?
I used crontab to set a file playing at a certain time, this works fine; however I want this to run even if no-one is logged in (but the computer is on). I can't get it to do this
Line is: 30 06 * * * env DISPLAY=:0.0 /usr/bin/totem /home/adunaic/Playlist.pls
Okay I think the problem was with needing a GUI.
Using this instead works, but i only hear the sound when I log in. I think something needs to be doen to start the audio perhaps?
I have this code that is 'bashed' regularly with crontab and basically it will send me an E-Mail of most of the output but it misses out some of it!
Here is the crontab code to automatically run the script:
Code:
So that sends me an E-mail with most of the output of the following code:
Code:
It sends me everything up to echo "*******" "Begin compressing and transferring files" "*******" but it wont output the tar bit.. so it should give me a list of files that have been tarred.
I am trying to have the files from /user/directory copy every hour to /backup/user/directory. It would seem that cron or crontab is what I need to use. Looking at previous posts and other documentation only shows how much I don't know. When I type crontab -e I get a blank file I can type into, seemingly using vi as the editor. I have no problem with that but when I type cron -l, I get my text after a message about "Do Not Edit this File". What I am reading just makes no sense, I am not understanding even the most fundamental aspects of cron or crontab. Where I can get the most basic of basic instructions to try to understand this function?
I have a question about using crontab with /etc/crontab...
I had a cron job that I needed to run as root. At the time I thought that sticking it in /etc/crontab would be a good idea. However, I used the crontab command to edit /etc/crontab, which I guess is not standard procedure? Specifically, I configured /etc/crontab as my local user's crontab (i.e. sudo crontab /etc/crontab) then added my cron job as I would a local user crontab (i.e. sudo crontab -e).
Originally, my cron job looked like this:
30 * * * * root /my/batch/script &> /dev/null
After adding the new cron job I started seeing errors. Something to the effect of "can't find command root" or something similar. So I removed the 'root' user definition from the cron job and the job started running fine. However, because this is /etc/crontab, there are other system related cron jobs that have been defined to run under the root account (e.g. "17 * * * * root cd / && run-parts --report /etc/cron.hourly" runs as root, etc.). So these pre-existing system cron jobs, which up until now have been running smoothly, are now generating "can't find command root" errors. But I think that the system cron jobs _are_ successfully being run someplace because logrotate seems to be working.
So what I _think_ is happening is that /etc/crontab is being run twice: once as the system crontab, and once as my sudoed local user's crontab. When I run crontab -l I see nothing, but when I run sudo crontab -l I can see the contents of /etc/crontab. I am reluctant to delete my sudoed local user's crontab, because then in the process I would be deleting the system crontab, and I do not know how I should restore the system crontab's contents. (I am still not sure as to the most appropriate way to edit the system crontab).
How can I get out of this mess? I want /etc/crontab to go back to the way it was before--running _once_ as the system crontab. As for my new cron job, I'm willing to reconfigure it anywhere so long as I am still able to run it as root. Any ideas? (I am using Ubuntu 8.04 Server LTE)
Is been a while since I was last here requesting help. Now I need some of that LQ magic. I have a script file that extracts data for a date range, create a zip file for the data and then ftp it to a remote server. Now the script is working perfectly if it is run manually, but now I want it to execute automatically. So I use the crontab, but for some reason the script fails when run from the crontab.code...
I am using squid 2.6 on my cent os 5 .I want to shutdown the system at 7:00 pm for that purpose I write in# crontab -e 0 19 * * * /sbin/shutdown -h nowThe thing which I want to clear that when the system will going to turn off it will also stop squid service? If not then how can I automatically stop squid service and then crontab file execute
when a script in /etc/cron.d directory will be executed?. I know that scripts in cron.daily will be executed daily [ set in /etc/crontab file]? Cant able to find this directory listed in /etc/crontab file?
I use the following code to send make-up file dialy via sftp with expect. When I run it from command line, it has no problem sending to the remoteSERVER side, but when it is running via crontab task, it did not do the put, so did not execute the batch file defined in -b option.
I plugged in my USb drive into my computer yesterday and tried to delete a folder. I was unable to do so and got the following message
Cannot move file to trash, do you want to delete immediately? The file "my file" cannot be moved to the trash. Show Details Unable to create trashing info file: Read-only file system
So when I click on delete I get another error message:
Error while deleting. There was an error deleting Case Study Database. Show Details Error removing file: Read-only file system
At this point I can only click on Skip, Skip All, or Cancel.
I have not changed anything on the stick recently so I dont know what is causing the problem.
Is there any program/script that can create a torrent file for each file in a directory? I have been looking all over but can't seem to find anything of the sort. I have 700+ files I REALLY don't want to make my self.
I want to create a logon script (or somesuch) that creates a file (if it doesn't already exist) and checks the file for some info otherwise. If it finds a given trigger in that file, it logs into a local database and does some operations.
Now my problem isn't with creating that file or even getting it to function as a logon script -- it's with permissions. After the logon script creates the file, I want that user to have read access on it ONLY. Further, I don't want to give the user any kind of root access so that they could do the database operations in question or chown/chmod the file.
What's the best practice here? I'm noticing that whenever the script runs (in .bashrc right now) the script runs with the current user's permissions. Ideally, I'd like to make it so the login script can run at a higher level of permissions, (higher than the user has). Is this even possible? What's the best way to do this?
I have iMac 2.4GHz with rEFIT installed. I installed Unity on one of the partition. Kernel is still 2.6.38.8. I am getting error message when I run the command sudo gedit /etc/x11/xorg.conf
The error message is
(gedit:2139): Gtk-WARNING **: Attempting to store changes into `/root/.local/share/recently-used.xbel', but failed: Failed to create file '/root/.local/share/recently-used.xbel.AC7YXV': No such file or directory
(gedit:2139): Gtk-WARNING **: Attempting to set the permissions of `/root/.local/share/recently-used.xbel', but failed: No such file or directory
I was just testing specifying limit on file size to a user and have added the following to /etc/security/limits.conf bob soft fsize 100 This basically should have said not to allow bob to create anyfile greater than 100Kb in size.
But the interesting thing is, if bob already has any file which is greater than 100Kb in size, it even doesn't allow to log him into the system both from console and SSH. Also nothing is logged in logs.. How do I configure it so that, bob can login to the system even though he has any file greater than 100Kb (but doesn't allow him to create file which are greater than 100Kb) ??
I want to back up an entire Linux system on a 3Tb external Western DIgital USB3 drive.
I do not want to reformat it from what it is, apparemtly NTFS.
Is there a utility that can act like a file manager like mc, that will permit me to create an ever expanding (to 320Gb) TAR file that will retain all the original file permissions. I have had nothing but disappointment with Linux backup utils with a FAT32 external drive, and I am concerned if I just try an tar the entire drive at once, with around 3 million files, I might run out of memory.