General :: Cron Job To Create And Move Backup From Ubuntu To Windows?
Jun 15, 2010
This is Kishore and i am new to Ubuntu and SVN and please some one help me in creating a cron job for my svn backup every day at 10:30 pm I already created a cron job which looks like 30 10 * * * svnadmin dump /home/administrator/svnrepository >svn1 when i run command directly i am getting whole backup and it's size is 3.6 gb but when i run through cron job the backup size is only 9 mb. So finally my requests are 1. cron job for taking complete svn backup at 10:30 pm daily and 2. cron job to copy the SVN backup in to my windows system in d drive and this must be run every day at 11:30 pm.
View 1 Replies
ADVERTISEMENT
Mar 3, 2011
How do you create a cron file that will regularly perform a level 0 backup once per month?
View 2 Replies
View Related
Sep 29, 2009
I have a php script in cron directory that generates 5 textfiles, after the files are generated, I want to create a script that will move the 5 text fiels to anoher folder name "web".
View 2 Replies
View Related
Jan 4, 2011
A user will be ftping some files to an upload directory. I need to move those files to another directory. I also need to mail a list of the just moved files to the user. This job will need to run every 10 minutes. I need to keep a log that holds all the files for the day that were moved, renaming it with the date/timestamp.I have this below but I just can't put it all together. a workable script out of this?
#!/bin/bash
# test.shl
source="/u03/upload/"
[code]...
View 5 Replies
View Related
Jun 8, 2009
Maybe this is a MySQL question, maybe not...
I've written a shell script to back up a database.
But when I run it, it prompts for password even though the script provides it. If I'm doing this manually, it's not a problem, but I want to make a cron job to do it...
Here's the script: Quote: #!/bin/bash
set -xv
#First let's rotate the backup files...
/bin/mv /home/cabazio/someDB-3.tar.gz /home/some/someDB-4.tar.gz
/bin/mv /home/some/someDB-2.tar.gz /home/some/someDB-3.tar.gz
[Code]....
View 2 Replies
View Related
Jan 5, 2011
I'm having a small issue where the backup jobs that I set to run in the crontab of the backup user do not appear to be running. Here's how I set it up (with crontab -e as the backup user):
run amanda every night (check at 2:45 and backup at 3)
[code]...
View 5 Replies
View Related
Oct 12, 2009
how to create cron file and how to run them?
View 2 Replies
View Related
Jul 30, 2011
Will this:
* * 1 1-12 * #dump -0uf /dev/st0/
* * * * 0 #dump -2uf /dev/st0/
* * 2-31 * 1-6 #dump -5uf /dev/st0/
answer this question?How to create a cron file that will regularly perform the following backups:a. Performs a level 0 backup once per monthb.Performs a level 2 dump one day per weekc.Performs a level 5 dump every day that neither a level 0 nor a level 2 dump is performed. If not, I'm I close? This is homework, so I don't want a direct answer
View 14 Replies
View Related
Nov 19, 2010
I am trying to clean my computer. Basically I try to order some movies to put all of them in directories. I started by hand but I am loosing time I guess. I want to do something like:
mv *.avi /nonexisting_directory/
nonexisting_directory being the name of the file without .avi, changing for each file. How can I write in command lineor in script?
View 10 Replies
View Related
May 4, 2010
i am very sorry if this has been asked before... i'm sure it has.. but i have searched all over the net looking for an answer and i still cant find it...
I have a really simple cron job script like this:
When i run this manually it works fine but when i run it from my ROOT user in Plesk as a cron task is always creates a file that is just 45 bytes. Why doesn't it work... I am running it as a root user.. so surely i must have permission to access the file?
View 7 Replies
View Related
Dec 2, 2010
I'm running out of disk space in root, so I decide to move /usr/lib64 directory to other partition by creating moving the whole directory to new location and create a softlink in the old location. What I'm planning to do is:
1. Copy the whole directory to new location
2. Make sure if there's no problem in the copy process
3. Create a softlink in the old location pointing to the new location.
However, when I'm comparing the old location and the new location using du command, there's some discrepancies between those two where the new location contain bigger size than the old one. How is this possible?
Is this the proper way to save disk space by creating a softlink?
View 5 Replies
View Related
Jun 10, 2010
I'm trying to set up a simple backup script with cron.
In "crontab -e" (and sudo crontab-e - I tried both) I enter "0 22 * * * /home/USERNAME/.backup.sh", with the hope that it will run the script at 10pm each day. The srcipt work fine if I run in a terminal. why it won't work? It's bound to be something obvious....
View 5 Replies
View Related
Jul 22, 2010
I'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet:
cd /media/"My Book"/"Linux Backups"
NOW=$(date +"%b-%d-%y")
LOGFILE=Backup_Root_FileSystem-$NOW.log
sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
[Code]...
View 5 Replies
View Related
Apr 3, 2010
I am trying to write as bash script in order to have backup done by cron on the webhosting server. I want all backup runs to have incremental number in front of them. I came up with an idea to store incremental number of a backup in txt file on a server so when next one runs is can check the number in the file. However I am having terrible issues. My script:
[code]....
View 7 Replies
View Related
Jul 20, 2009
An external service I dont manage pushes mediafiles into a shared directory on my server. I need to move these files into their correct directories automatically. The problem is that if I run my script as a cronjob once every 3 minutes, I notice that the script copies files which are still on their way into my server. So I need to figure out how to have the script check that the files are complete (done downloading) before the script moves the files. This is what I got so far:
Code:
#!/bin/sh
# Script by proximity 280709.
# Locate correct directory
[code]....
View 8 Replies
View Related
Jul 5, 2010
In doing my kernel-kits that allow you to create a livecd backup of your installed slackware or arch install, I came across stuff about how many x86 or x86_64 systems use a PAE kernel?
if thats so; should I release a PAE kernel kit also besides the x86 and x86_64 kernel kits for slackware/arch??
or is it not a big thing?
Can anyone run a PAE kernel or it should be a seperate build? the x86/x86_64 kits are not PAE...let me check...
View 4 Replies
View Related
Dec 7, 2010
I Acer aspire 3620 and would like to backup whole system. My laptop has a 40Gig hdd. Is there a way I can either create an image or copy(clone) to another computer just in case I need it. If it matters I have a spare 40 gig drive on the other computer. The reason for doing this is so that I can try a system restore and if anything goes wrong I want to be able to transfer back. Also if it is posible how could I put it back and be bootable.
View 14 Replies
View Related
Dec 23, 2010
I'm running a cron job every night to dump a MySQL database to an external hard drive. It works, however when I check on it the following morning the external is no longer mounted and the XFS log file is corrupted. If I run
Code:
xfs_repair -L /dev/sdf1
It works, but then I get these issues:
Code:
XFS: Filesystem sdf1 has duplicate UUID - can't mount
I can reset the UUID, but it's difficult to have to do this every day.
View 2 Replies
View Related
Oct 24, 2009
What's a good cron script for backing up and zipping a directory of files, or multiple directories with files, to a backup directory on my server, on a daily basis?I found an easy to use mysql backup script, now I need to backup my site directory, but not all the directories in it. So I need a method in the script to omit certain directories from backing up, ie dirs that contain gigs worth of files.This seems like it should be one of the most common crons to set a server up with but two pages deep in google (and here) I have yet to find anything remotely resembling a solution.
View 9 Replies
View Related
Sep 20, 2010
I did a search and came up with many different opinions and suggestions, and I could not find any similar threads for this (if there is I must apoligize in advance and did not look hard enough). I am trying to create a backup schedule for client computers and servers; and I would like your opinion on what I came up with. Here is network infrastrusture setup:
16 client computers
6 serevrs
- Primary and Secondary Web Server (RAID 1)
- Primary and Secondary Domain Controller (RAID 1)
- File/Print Server (RAID 5)
- Backup Server (RAID 1) with USB 2.0 1TB External Drive
Of course I will evetually upgrade to Tape Drives, but I am saving up for that. Anyways I want to backup ALL client computers, Primary Web and DC server, and File/Print Server. This will be done (of course) on the Backup Server) Here is the schedule:
[Code]..
View 2 Replies
View Related
Oct 15, 2010
Due to a disk crash I've had to rebuild my Debian Lenny system. For some reason I can't get my cron-fired backup scripts to run. They will run manually.
It looks like crond is not running. If I try to start it, here's what I get:
Pancho:/home/lloyd# /etc/init.d/cron start
Starting periodic command scheduler: crond failed!
MORE INFO:
lloyd@Pancho:~$ /etc/init.d/cron start
Starting periodic command scheduler: crond/etc/init.d/cron: line 54: start-stop-daemon: command not found
failed!
[Code].....
Clearly the problem failure to find start-stop-daemon is not the problem and I'm still in the dark.
View 2 Replies
View Related
Feb 9, 2010
we have a server that runs a backup (cron job) at 9:15 every night. When I log on in the morning I have mail message that gives me a long list of all the files that were backed up the night before. For a couple of weeks now, the mail message gives me an empty list. Yet, when I run the same job manually from a #prompt, it runs. I am not able to run this job with cron in the daytime because too many users are in it. I wanted to browse the tape to see if the backup is really failing to copy the files or if they are on the tape and the mail message is bogus.
Since the backkup was done with cpio instead of tar, I'm not sure if I can browse the tape with restore -i anyway.What would be the best way to browse the tape on /dev/rmt/1 without actually restoring anything ?This is an ancient DGUX system, not Linux, and I'm not a unix expert I just inherited this server recently, but a lot of things are very similar to Linux and it looked like this might be a good place to ask.
View 3 Replies
View Related
Mar 2, 2011
I own a CentOS 5 VPS. I typed crontab -e, and then I added the following line to automatically have my server backup mysql
0 * * * * mysqldump -u root -p password --all-databases | gzip > /home/dbbackup/database_`date '+%m-%d-%Y_%H'`.sql.gz
When I go in and look, it doesn't place any files in /home/dbbackup. When I run
mysqldump -u root -p password --all-databases | gzip > /home/dbbackup/database_`date '+%m-%d-%Y_%H'`.sql.gz
View 3 Replies
View Related
Jun 13, 2010
So I've been trying to solve another problem (disk errors that MIGHT have something to do with ext4). I was wondering if there would be a possible way to backup my 10.04 install and move it to an existing, unused ext3 partition. I'm kinda stuck at the moment with my disk error problem, and figured I could waste some time to see if I could MOVE my install (as to not have to do a fresh install, re-install every application and script, and re-customizeand and move all of my media and files) onto an ext3 and completely remove the ext4.
View 3 Replies
View Related
Aug 28, 2010
I have a cron backup scheme in which I rsync, then tar, then copy files on my internal hard drives to an external (USB) drive. When it works, it works. But I often get a "Permission Denied" message for all of these tasks. how the external drive is auto-mounted so I edited the etc/fstab so that the owner of the cron job is also the owner of the external drive (I think. Unfortunately, I'm not at that machine right now (it's at work), I can't give the exact fstab line (I will post it as an update to this thread next time I am at the machine).) BUT, I still get times when the cron backup runs fine and other times I get the Permission Denied. This is a shared machine that is dual-booted, so what I *think* is going on is that when the machine is rebooted to Fedora, but nobody logs in, I get a Permission Denied for the cron backup. It seems like on days when someone has logged in as the main user and left without logging out that the cron backup runs fine.
View 8 Replies
View Related
Jan 4, 2010
I installed a second HD, and formatted it to ext4. I gave it the "/backup" label. I am trying to figure out how to mount it so that I can run cron to backup my home folder onto it once a week. This is what the fstab looks like now
[code]...
View 9 Replies
View Related
Oct 7, 2010
How to create cron tab when DSL line down set automatically restart the network service while DSL line up?
View 3 Replies
View Related
Apr 11, 2011
I'm trying to move from Windows to Linux, and one of the main things I need is working television. I'd like to ask: Is it possible? If so, what kinds of Linux can deal best with this? What exactly should I do? Should a newbie like me try this at all? I don't know if it helps, but my TV card is KWorld Global TV Terminator PVR-7131RF and the program I used to use for this on Windows was PVR Plus
View 3 Replies
View Related
Aug 14, 2011
I would like to backup important files (totaling about 400GB) on my ext 4 RAID 5 array to an ext4 external hard drive over USB (external drive is mounted to /mnt. In the future I'd like to automate the process using rsync and cron so for now I'm using rsync to transfer the files. My problem is that using the rsync command like this: # rsync -Pr "/dir1" "/dir2" "/dir3" "/dir4" /mnt
rsync shows me the checks and transfers for awhile and then throws up an i/o error (wish I had a screenshot to show but I don't). When I ls /mnt I get a similar i/o error. I then check /dev for the drive and find that it no longer shows up. Originally the partition was /dev/sdc1. I tried unplugging the USB at this point, plugging it back in and mounting the drive back to /mnt, however it has now assigned it to (you guessed it) /dev/sdd1. I get the drive mounted and try the original rsync command again, hoping the first error was a fluke or some kind of one-time drive fart. This time it makes it quite a bit further and then throws up the exact same problem. Am I doing something terribly wrong here? As I said, I'm very new to bash so I'm not making some absolutely moronic, newbie mistake.
View 9 Replies
View Related
Mar 13, 2010
I am trying to use rsync & ssh to move a backup folder some computers to a server. I found a command that is supposed to do this, but I am having issues getting it to work.
Code:
rsync -avz -e 'ssh -p 'port' 'username'@'hostname' ' /source/path host:/destination/path
View 2 Replies
View Related