Debian Configuration :: Can't Get Cron-fired Backup Scripts To Run
Oct 15, 2010
Due to a disk crash I've had to rebuild my Debian Lenny system. For some reason I can't get my cron-fired backup scripts to run. They will run manually.
It looks like crond is not running. If I try to start it, here's what I get:
lloyd@Pancho:~$ /etc/init.d/cron start
Starting periodic command scheduler: crond/etc/init.d/cron: line 54: start-stop-daemon: command not found
failed!
[Code].....
Clearly the problem failure to find start-stop-daemon is not the problem and I'm still in the dark.
Just curious if this is possible. What I want to do is setup a rsync job to backup my laptop to my personal file server(same LAN), but I want it do do this at 3AM while I'm sleeping. However I always close my laptop lid when I'm done using it which puts the laptop into suspend mode since that's how I configured my power options in gnome. Since I don't want to leave my laptop powered up all the time, I was wondering if it is possible to have a cron job scheduled that will wake the laptop up(out of suspend mode) and run my script/backup job, all without opening the lid of the laptop, and then put it back into suspend mode when it's done. Is this possible, and is it as easy as scheduling a cron job or is there some other scripting/configuration/trickery that I need to do to accomplish this?
Also, my laptop's BIOS has the option to power on at a scheduled time if needed, but I'm not sure if that would work with it being suspended(hibernation is not an option since my entire HD is encrypted with LUKS and I would have to be present to enter the password to boot the system.).
I have a XEN DomU VM running Lenny and set up using debootstrap. I also have the TimeWentBackwards issue sorted using: [URL]. The problem is that my cron jobs just don't run. I've set up a cron job that simply echoes a date string to a file as a tester but it's not running. I've tried it under the root crontab, in /etc/crontab and set up in /etc/cron.daily. Is there something I'm missing?
The system crontab (/etc/crontab) uses the same format, except that the username for the command is specified after the time and date fields and before the command.
[code]...
Every hour, I get an e-mail complaining about the first line of the crontab:
[code]...
I get the same complaint from the other entries: It looks to me as if cron, or anacron, is trying to execute the user (root) as a command. Predictably, the shell doesn't like it, so barfs and triggers an e-mail about it. Why is this not doing what the man page says it should do? The 2nd problem I believe is related to exim, not cron. The e-mails I'm getting above are being bounced from my ISP because they are directed to root@myisp.com, rather than my regular e-mail address. When the message bounces, it bounces to my regular e-mail address. In /etc/aliases, I have root: [URL]... and in etc/email-addresses I have root: [URL]... Adding the entry to /etc/email-addresses allowed the bounce to find me because the sender's address is [URL]... but how can I get cron to send these messages to me in the first place, instead of root?
I am making backups and I need to make a cron job that mounts a 2nd local hard drive.
It is not listed in my fstab file and I mount it manually in nautilus (having to type a password). It is designated as /dev/sdb1 and /media/repo when it is mounted. Can I get cron to mount it and then add the password or do I have to add it to fstab?
Since we switched our server to Squeeze, I'm receiving emails from the cron about a mysql error (...can't use locks with log tables). But this is not the issue I want to talk about here. The issue is that since the executed command is: /usr/bin/mysqlcheck -uroot -pmypassword --all-databases --check-only-changed --silent which is a command that was automatically added to the cron by the mysql package (I believe), the password is sent in clear text in the email's subject. In my sense, this is a serious security issue (sending root password in email subject...)
I don't know at what level it should be corrected, but it seems to me like it should be corrected in the Debian distrib, shouldn't it ?
And for now, how can I hide the password in the emails I receive ?
I've been following grub-common bug #606845 and in coming to a solution for the issue, these guys are using dd as a brute force means of swapping out master boot records or trampling them, if you prefer. (Background: The issue is related to grub and certain xp installations)
An sample snip of code: dd if=/mbrxp.bin of=/dev/sda bs=512 count=1 1) Is mbrxp.bin a back-up of the mbr taken before installation of squeeze (or grub in general)? 2) Am I fubar if I didn't make a back-up of the mbr before installing grub?
Is it possible (on lenny, in case that matters) to backup some directories with rdiff-backup, with the target being a WinXP Pro Host, i.e. the target being a SMB share? My idea is to start the XP-Box over WOL, run rdiff-backup and then shut it down using "net rpc SHUTDOWN"...
How to remove a specific folder from your backup?$ rdiff-backup --remove-older-than now /backup/backup_laptop/home/derick/DownloadsFatal Error: Increments for directory /backup/backup_laptop/home/derick/Downloads cannot be removed separately.Instead run on entire directory /backup/backup_laptop.
I've written a shell script to back up a database.
But when I run it, it prompts for password even though the script provides it. If I'm doing this manually, it's not a problem, but I want to make a cron job to do it...
Here's the script: Quote: #!/bin/bash set -xv #First let's rotate the backup files... /bin/mv /home/cabazio/someDB-3.tar.gz /home/some/someDB-4.tar.gz /bin/mv /home/some/someDB-2.tar.gz /home/some/someDB-3.tar.gz
i am very sorry if this has been asked before... i'm sure it has.. but i have searched all over the net looking for an answer and i still cant find it...
I have a really simple cron job script like this:
When i run this manually it works fine but when i run it from my ROOT user in Plesk as a cron task is always creates a file that is just 45 bytes. Why doesn't it work... I am running it as a root user.. so surely i must have permission to access the file?
I'm having a small issue where the backup jobs that I set to run in the crontab of the backup user do not appear to be running. Here's how I set it up (with crontab -e as the backup user):
run amanda every night (check at 2:45 and backup at 3)
I'm trying to set up a simple backup script with cron.
In "crontab -e" (and sudo crontab-e - I tried both) I enter "0 22 * * * /home/USERNAME/.backup.sh", with the hope that it will run the script at 10pm each day. The srcipt work fine if I run in a terminal. why it won't work? It's bound to be something obvious....
What's a good cron script for backing up and zipping a directory of files, or multiple directories with files, to a backup directory on my server, on a daily basis?I found an easy to use mysql backup script, now I need to backup my site directory, but not all the directories in it. So I need a method in the script to omit certain directories from backing up, ie dirs that contain gigs worth of files.This seems like it should be one of the most common crons to set a server up with but two pages deep in google (and here) I have yet to find anything remotely resembling a solution.
I am running on a laptop and cron.daily is set to run at 0625 So I wonder what happens if my machine is not turned on at that time.. At that rate it could also be off for the other periods as well (weekly, monthly) Is there solution that will allow them to run once they are online after the appointed time? using a cron entry that runs every 15 or 5 or 1 minute.
I am trying to write as bash script in order to have backup done by cron on the webhosting server. I want all backup runs to have incremental number in front of them. I came up with an idea to store incremental number of a backup in txt file on a server so when next one runs is can check the number in the file. However I am having terrible issues. My script:
we have a server that runs a backup (cron job) at 9:15 every night. When I log on in the morning I have mail message that gives me a long list of all the files that were backed up the night before. For a couple of weeks now, the mail message gives me an empty list. Yet, when I run the same job manually from a #prompt, it runs. I am not able to run this job with cron in the daytime because too many users are in it. I wanted to browse the tape to see if the backup is really failing to copy the files or if they are on the tape and the mail message is bogus.
Since the backkup was done with cpio instead of tar, I'm not sure if I can browse the tape with restore -i anyway.What would be the best way to browse the tape on /dev/rmt/1 without actually restoring anything ?This is an ancient DGUX system, not Linux, and I'm not a unix expert I just inherited this server recently, but a lot of things are very similar to Linux and it looked like this might be a good place to ask.
I have a cron backup scheme in which I rsync, then tar, then copy files on my internal hard drives to an external (USB) drive. When it works, it works. But I often get a "Permission Denied" message for all of these tasks. how the external drive is auto-mounted so I edited the etc/fstab so that the owner of the cron job is also the owner of the external drive (I think. Unfortunately, I'm not at that machine right now (it's at work), I can't give the exact fstab line (I will post it as an update to this thread next time I am at the machine).) BUT, I still get times when the cron backup runs fine and other times I get the Permission Denied. This is a shared machine that is dual-booted, so what I *think* is going on is that when the machine is rebooted to Fedora, but nobody logs in, I get a Permission Denied for the cron backup. It seems like on days when someone has logged in as the main user and left without logging out that the cron backup runs fine.
I installed a second HD, and formatted it to ext4. I gave it the "/backup" label. I am trying to figure out how to mount it so that I can run cron to backup my home folder onto it once a week. This is what the fstab looks like now
I'm running a cron job every night to dump a MySQL database to an external hard drive. It works, however when I check on it the following morning the external is no longer mounted and the XFS log file is corrupted. If I run
Code: xfs_repair -L /dev/sdf1 It works, but then I get these issues: Code: XFS: Filesystem sdf1 has duplicate UUID - can't mount I can reset the UUID, but it's difficult to have to do this every day.
This is Kishore and i am new to Ubuntu and SVN and please some one help me in creating a cron job for my svn backup every day at 10:30 pm I already created a cron job which looks like 30 10 * * * svnadmin dump /home/administrator/svnrepository >svn1 when i run command directly i am getting whole backup and it's size is 3.6 gb but when i run through cron job the backup size is only 9 mb. So finally my requests are 1. cron job for taking complete svn backup at 10:30 pm daily and 2. cron job to copy the SVN backup in to my windows system in d drive and this must be run every day at 11:30 pm.
I would like to backup important files (totaling about 400GB) on my ext 4 RAID 5 array to an ext4 external hard drive over USB (external drive is mounted to /mnt. In the future I'd like to automate the process using rsync and cron so for now I'm using rsync to transfer the files. My problem is that using the rsync command like this: # rsync -Pr "/dir1" "/dir2" "/dir3" "/dir4" /mnt
rsync shows me the checks and transfers for awhile and then throws up an i/o error (wish I had a screenshot to show but I don't). When I ls /mnt I get a similar i/o error. I then check /dev for the drive and find that it no longer shows up. Originally the partition was /dev/sdc1. I tried unplugging the USB at this point, plugging it back in and mounting the drive back to /mnt, however it has now assigned it to (you guessed it) /dev/sdd1. I get the drive mounted and try the original rsync command again, hoping the first error was a fluke or some kind of one-time drive fart. This time it makes it quite a bit further and then throws up the exact same problem. Am I doing something terribly wrong here? As I said, I'm very new to bash so I'm not making some absolutely moronic, newbie mistake.
Can anybody help me out with keystroke logging. I am trying to figure out if there is any way to get all the keystrokes fired to my laptop in my absense. Is there any packages for this ?
I want to edit files while in runlevel three. I am trying to learn how to navigate my entire system via CL. Whenever I use gedit in runlevel three, is says, "cannot open display"
Is there a utility to edit files without X11 fired up?
Is there any utility which will monitor all commands on SSH SHELL fired by any user who logs in using putty or any other client? If this record is saved somewhere,
This script simply deletes files older than a certain age (in this case 7 days) from a certain location; I use it to purge old backups nightly, and it works as expected:
# delete backups older than 7 days find /mnt/backup/* -mtime +7 -exec rm -Rf {} ;
The problem is, every morning I get an email with an error message something like this:
find: `/mnt/backup/subfolder': No such file or directory
I am now preparing myself to upgrade lenny to squeeze and decided to do a backup on my system. I used backup-manager to do the job and it worked fine. how do you restore said backup data?
1. Log in as the root user. 2. I have created a file with name "reminder" in /root directory. 3. Create a /etc/cron.daily directory. Add a file called "taxrem", which reads a text file from home directory, so write a command in the "taxrem" : "cat ~/reminder" 4. Add command to /etc/crontab file. Based on the conditions I want, such as : 5 13 * * * root run-parts /etc/cron.daily
There is no entry in the cron.deny file. Still I have not get any response on that scheduled time.