General :: Create A Cron File That Will Regularly Perform A Level 0 Backup Once Per Month?
Mar 3, 2011How do you create a cron file that will regularly perform a level 0 backup once per month?
View 2 RepliesHow do you create a cron file that will regularly perform a level 0 backup once per month?
View 2 RepliesWill this:
* * 1 1-12 * #dump -0uf /dev/st0/
* * * * 0 #dump -2uf /dev/st0/
* * 2-31 * 1-6 #dump -5uf /dev/st0/
answer this question?How to create a cron file that will regularly perform the following backups:a. Performs a level 0 backup once per monthb.Performs a level 2 dump one day per weekc.Performs a level 5 dump every day that neither a level 0 nor a level 2 dump is performed. If not, I'm I close? This is homework, so I don't want a direct answer
This is Kishore and i am new to Ubuntu and SVN and please some one help me in creating a cron job for my svn backup every day at 10:30 pm I already created a cron job which looks like 30 10 * * * svnadmin dump /home/administrator/svnrepository >svn1 when i run command directly i am getting whole backup and it's size is 3.6 gb but when i run through cron job the backup size is only 9 mb. So finally my requests are 1. cron job for taking complete svn backup at 10:30 pm daily and 2. cron job to copy the SVN backup in to my windows system in d drive and this must be run every day at 11:30 pm.
View 1 Replies View RelatedHow to create a cron file that will regularly perform the following backups:
a. Performs a level 0 backup once per month
b. Performs a level 2 dump one day per week
c. Performs a level 5 dump every day that neither a level 0 nor a level 2 dump is performed.
In the worst case, how many restores would you have to perform to recover a file that was dumped using this schedule.
There must be a way to perform privileged tasks in the Gnome desktop (or actually Nautilus, I guess)...similar to using SUDO in a terminal window.
For example I want to change the rights to a file, but Nautilus tells me I can't because I'm not the owner.
I have a couple of Lenny LAMP servers, and a backup server. (virtual testing environment)
1. What is the best way to perform a backup? (system state as well as individual files) Although system state can also be accomplished through the hypervisor.
2. Between Windows computers, I access shared directories simply by \hostnamesharedmap or \host_IPsharedmap. Between Windows and linux i use SAMBA. But there must be a simple way to copy 2 files between linux hosts?
3. I've searched a lot, and only found people with the same question without a good answer: is there a linux equivalent for robocopy?
I need to use Live Media to make changes to a hard disk drive. I tried to sudo to no effect.
Seems like it would be troublesome to authorize root access on Live Media or have sudo allow commands such as pvcreate,lvcreate,vgcreate, especially if the hard disk is not encrypted.
Firstly i'd like to introduce myself before I ask my question, i am currently studying a "Network Operating System 2" course, and in the syllabus the professor decided to use Ubuntu Linux for us. This is a first for me and I am truly greatly interested in the OS. It seems like a solid system once you get to know the deep foundations of it. The professor has already given us an assignment, but I feel its not fair as its just a bit too much from out of what he taught us .. He did not teach us what he is asking for. He asked us to create to a text file which lists the Sales from two Salespersons, and each one will have an amount for each month .. So 12 entries for each salesperson.
Using that text file, we should write a bash script which would read the values from the above text file, and produce a "Monthly Sales" text file, which shows the total sales of each month (combined from both the salespersons). I would like to ask for some guidance, some help on how to start. It's just that he didn't teach us any commands about this matter.
i want backup from my system fdora11 64bit .at that i worked on this command, rsync -avz soursepath /media/destination path but it is not working correctly
how can i take backup to my external hard disk
I have a samba file server that has Windows XP machines connecting to it. Fairly regularly, I get messages such as the following. Are these anything to worry about? What do they mean? Nobody ever has problems with disconnecting or losing their mapped drive or their desktop shortcut not working.
Code:
Nov 15 15:37:21 servername smbd[27516]: [2010/11/15 15:37:21, 0] lib/util_sock.c:read_socket_data(384)
Nov 15 15:37:21 servername smbd[27516]: read_socket_data: recv failure for 4. Error = No route to host
[code]....
Maybe this is a MySQL question, maybe not...
I've written a shell script to back up a database.
But when I run it, it prompts for password even though the script provides it. If I'm doing this manually, it's not a problem, but I want to make a cron job to do it...
Here's the script: Quote: #!/bin/bash
set -xv
#First let's rotate the backup files...
/bin/mv /home/cabazio/someDB-3.tar.gz /home/some/someDB-4.tar.gz
/bin/mv /home/some/someDB-2.tar.gz /home/some/someDB-3.tar.gz
[Code]....
I'm having a small issue where the backup jobs that I set to run in the crontab of the backup user do not appear to be running. Here's how I set it up (with crontab -e as the backup user):
run amanda every night (check at 2:45 and backup at 3)
[code]...
I want to back up an entire Linux system on a 3Tb external Western DIgital USB3 drive.
I do not want to reformat it from what it is, apparemtly NTFS.
Is there a utility that can act like a file manager like mc, that will permit me to create an ever expanding (to 320Gb) TAR file that will retain all the original file permissions. I have had nothing but disappointment with Linux backup utils with a FAT32 external drive, and I am concerned if I just try an tar the entire drive at once, with around 3 million files, I might run out of memory.
I am a backup noob. My idea of backing something up is finding a big enough flash drive and copying the necessary files over.
So I really need to learn now. I'm wiping a Vista laptop for a friend to install Windows 7. But first, I want to do a whole-drive backup in case something goes wrong. It's a 100GB drive with 50GB of data.
Is it possible that I could do this via my home network or via a direct ethernet connection? I have a desktop with a 1TB drive I could back up to. Like I say, I'm a noob so I'm open to anything.
One more thing: I'd like this backup to be in a form that I can retrieve individual files from it if necessary. If everything goes right, I'll probably want to pull My Documents out of the backup and drop it into Windows 7.
Oh, and why am I asking on UbuntuForums instead of a Windows forum? Because I'm betting I'll end up booting a live CD on the laptop to do the backup. But I'm just guessing. At any rate, I'm sure I'll use Ubuntu tools, because that's what I know.
how i can use the cal command so it displays the current month and the next month, i know you can use cal -3 9 2007 to show the previous current and next but i just want to display the current and next.
View 9 Replies View RelatedI'm looking for a powerful network traffic monitor that can do all of the following (or at least a combination of tools that can do the following):
Tell me how much data was downloaded/uploaded on an interface this month and the previous month tell me how the traffic was used throughout the monthshow which internal IPs (IPs in the 192.168.1.0/24 network) used how much traffic show which ports/protocols on those IPs used all that traffic
Hhow LIVE traffic flow statistics that can tell me total speed of traffic going through an interface as well asshow which internal IPs (IPs in the 192.168.1.0/24 network) are using how much of the traffic show which ports/protocols on those IPs are using that traffic
This tool will run on a linux router through which all my internal PCs are connected to the Internet. This means the tool(s) need to work with NAT (traffic being forwarded and not necessarily destined for the interfaced being monitored).
The distribution being run doesn't have a package manager so any packages or dependencies have to be manually compiled and SCPed over file by file. For this reason, the tool/tools need to be simple (things like vnstat, not things like ntop that have their own web interface).
I know that vnstat can tell me the first bullet point so it's only there incase there's a tool out there that can do everything. If there's a tool that can only do the second or third bullet point, that's great too - I'll just keep using vnstat and look for something else to do the other task.
I would like to create a full systembackup to a ISO/IMG-file. I've been searching and found mondorescue.org, but something is wrong with package for debian 6.
View 14 Replies View Relatedhow to create cron file and how to run them?
View 2 Replies View RelatedI launch a mdadm grow to expand my filesystem but I forget --backup-file.I have my server running and one UPS.
Code: Select allpk25.com:~# mdadm --manage /dev/md0 --add /dev/sde1
mdadm: added /dev/sde1
pk25.com:~# mdadm --grow --raid-devices=5 /dev/md0
mdadm: Need to backup 3072K of critical section..
pk25.com:~# mdadm --grow --raid-devices=5 --backup-file=/root/md0-grow.bak /dev/md0
mdadm: /dev/md0 is performing resync/recovery and cannot be reshaped
[code]....
Attempting to create a backup script to copy files from one file system to a remote file system.
When I try this I get:
Quote:
# tar -cf - /mnt/raid_md1 | gzip -c | ssh -i ~/.ssh/key -l user@192.168.1.1 "cat > /mnt/backup/fileserver.md1.tar.gz"
tar: Removing leading `/' from member names
Pseudo-terminal will not be allocated because stdin is not a terminal.
ssh: Could not resolve hostname cat > /mnt/backup/fileserver.md1.tar.gz: Name or service not known
[Code].....
I know that the remote file system dir is RW and the access is working fine. I am stumped...
I have a folder named "logs" which contains log files. I would like to have a bash script or a cron job or something like this which will check the folder continuousy and perform some tasks (for example echo "New file is created") every time a new file with pattern like "Screen.log.***" is created in the folder "logs"
Is it possible in Linux?
Can some one give me a sample of a crontab for backing a directory please, System is Ubuntu 9.04Quote:
#!/bin/bash
# this file is an automated backup script, backup.sh.
# this backs up my domain site.
[code]....
i am very sorry if this has been asked before... i'm sure it has.. but i have searched all over the net looking for an answer and i still cant find it...
I have a really simple cron job script like this:
When i run this manually it works fine but when i run it from my ROOT user in Plesk as a cron task is always creates a file that is just 45 bytes. Why doesn't it work... I am running it as a root user.. so surely i must have permission to access the file?
In doing my kernel-kits that allow you to create a livecd backup of your installed slackware or arch install, I came across stuff about how many x86 or x86_64 systems use a PAE kernel?
if thats so; should I release a PAE kernel kit also besides the x86 and x86_64 kernel kits for slackware/arch??
or is it not a big thing?
Can anyone run a PAE kernel or it should be a seperate build? the x86/x86_64 kits are not PAE...let me check...
I Acer aspire 3620 and would like to backup whole system. My laptop has a 40Gig hdd. Is there a way I can either create an image or copy(clone) to another computer just in case I need it. If it matters I have a spare 40 gig drive on the other computer. The reason for doing this is so that I can try a system restore and if anything goes wrong I want to be able to transfer back. Also if it is posible how could I put it back and be bootable.
View 14 Replies View Relatedhello everyone, im having a problem when my computer enters in the run level 4 as the default when i start slackware. The strange thing is that it not seems that is a X window problem, it looks like more like a configuration problem in some part of the kde script to initialize the log in, because if i manually start the X service it works fine, i dont know what is the source. Thank you in advance for the help.
View 3 Replies View RelatedI'm trying to set up a simple backup script with cron.
In "crontab -e" (and sudo crontab-e - I tried both) I enter "0 22 * * * /home/USERNAME/.backup.sh", with the hope that it will run the script at 10pm each day. The srcipt work fine if I run in a terminal. why it won't work? It's bound to be something obvious....
What's a good cron script for backing up and zipping a directory of files, or multiple directories with files, to a backup directory on my server, on a daily basis?I found an easy to use mysql backup script, now I need to backup my site directory, but not all the directories in it. So I need a method in the script to omit certain directories from backing up, ie dirs that contain gigs worth of files.This seems like it should be one of the most common crons to set a server up with but two pages deep in google (and here) I have yet to find anything remotely resembling a solution.
View 9 Replies View RelatedI'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet:
cd /media/"My Book"/"Linux Backups"
NOW=$(date +"%b-%d-%y")
LOGFILE=Backup_Root_FileSystem-$NOW.log
sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
[Code]...
I did a search and came up with many different opinions and suggestions, and I could not find any similar threads for this (if there is I must apoligize in advance and did not look hard enough). I am trying to create a backup schedule for client computers and servers; and I would like your opinion on what I came up with. Here is network infrastrusture setup:
16 client computers
6 serevrs
- Primary and Secondary Web Server (RAID 1)
- Primary and Secondary Domain Controller (RAID 1)
- File/Print Server (RAID 5)
- Backup Server (RAID 1) with USB 2.0 1TB External Drive
Of course I will evetually upgrade to Tape Drives, but I am saving up for that. Anyways I want to backup ALL client computers, Primary Web and DC server, and File/Print Server. This will be done (of course) on the Backup Server) Here is the schedule:
[Code]..