I just got a 2TB drive with the intention of backing up multiple Ubuntu machines to it. What would be the best way to do this, keeping ease of restoration in mind? Should I just copy each drive image to the BU drive, or use a utility like Back in Time?
New to Ubuntu (and Linux in general) and have whipped up a quick python script to backup my files and do some other housekeeping. I don't see any service scheduling tools and was wondering what the best way to do this is.
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
I have been a windows user for a very very long time, and recently switched to debian. I have so to speak crapped all over my system failing to compile sourcecode properly held broken repos and what not.
So decided to reinstall debian last night, now I have a rather clean fresh installed copy. As I'm a new user to linux, I want to setup a backup system so I can revert to a clean fresh installation when I muck up my system.
Now I have downloaded and installed BackupPC with needed libraries and sitting here trying to configure it on my computer. I have followed this guide: [URL] ....
I have come to this section here:
********************** Server SSH KEY Creation and Deployment:
3. Deploy Key To Client Machine Copy id_rsa.pub to client machine
Now I'm not quite sure what to do here as I have no networked computers to deploy to only one machine. Although I want to make a local system backup and upload that to a server I have access to, but would that server be seen as a backuppc server/client?
On this computer here I have generated the id_rsa and id_rsa.pub located in /var/lib/backuppc/.ssh
looking at this part: Test ssh connection(On server, as Linux user backuppc) backuppc@server$ ssh root@<client machine>
How should I interpret @<client machine> should that to be ssh email@example.com ?
I have a machine running Karmic sat in the corner of my office. I scp stuff from my mac to it quite a lot over my local wireless network.At the moment, it holds a .sparsebundle backup of my mac created by SuperDuper! (mac backup application) that I periodically transfer over. I'd like to make this system a little more sophisticated and use Apple's Time Machine to automatically backup to my Ubuntu box if possible.
Anyone have any experience of this? I'm kinda hoping for something a bit more straight forward than this: http:[url]....
I'm running an Ubuntu 9.10 Linux server. I'm trying to find a way to backup the machine while it is running and from what I see, this eliminates the disk clone utilities. All of the disk clone stuff I have seen for Linux requires that you reboot into a special live CD.So my question is this, what is the best solution for backing up the system while it is running? Also, I don't really care about the OS config too much, I just want to be able to keep my stored files and my programs that I have installed on it.
I am using MySQL as the database system for my application on a Linux system. Every week I update the system and take backups (mysqldump) of the databases changed (2 databases). I then .tar.gz them and ftp the resulting file to a remote server, after which I remove the original backups and tar.gz files from the Linux server. Being a complete novice when it comes to Unix systems, I would like to know if it is possible to write a script which would do all this automatically, i.e. perform the following steps.
1) Backup database A to A.sql (mysqldump) 2) Backup database B to B.sql (mysqldump) 3) tar -cvzf dest.tar.gz A.sql B.sql 4) ftp dest.tar.gz to firstname.lastname@example.org 5) Delete A.sql, B.sql, dest.tar from local server
I have 2 servers. One is an operational CentOS + cPanel server and the other is a blank CentOS server. I want to use cPanel's backup suite to backup our customers accounts but I want to do it so the blank server is mounted to the cPanel server, rather than an FTP backup. If that makes sense. I believe from previous experience that NFS is the way to go but I'm not sure.
I am looking for a piece of hardware that would be able to run Linux and act as a web, subversion and file server. Ideally it shouldn't have any fans, because it will be in the middle of my living room, the computing power needs are minimal. I would be grateful for recommendations.
This is my first post and I am a linux newbie.. took on the challenge of setting up an ubuntu server with proxy and firewall with vpn access as well.. sounds good? While messing with things, I installed ebox server, know called Zentyal and perform backups, however, when performing a full restore, following instructions and all it does not restore correctly. I ran many attempts with verious configurations and still allways got some errors, such as eboc-ebackup failing to start, etc. most noteably wa sthe ldap error, which i figured out how to fix by restarting once logged into and then rebooting.
I say the Clonezilla options for a full backup, and it sounds great, but does require me to perform a full abckup, while bring the server down to perform. What I wanted to know, was what do you recommend for performing a full backup (possibly to NTFS partition or USB key with FAT32) with automation or even if i had to perform the trigger, just without needing to bring the server down, so when i go into production i dont have any issues backing up anytime?
I cant update php in my RHEL Server 5.1 using yum. I am using RHEL Server 5.1 and I cant use yum to upgrade my php. Other site told me that use wget to solve this problem. How to use wget to upgrade php? This is my first time to handle linux server..
I am trying to create a backup script that will back up a single folder for a class i am in. I was wandering if I could get some help. If possible I would also like to know how to write a script that can encrypt that same file . I will be putting the back up in my /home/usr/Backup directory. I am not trying to back up my whole system just a single folder. I am using Fedora 11
I am administering a live web server i want to keep a backup of the access log file without disturbing the server performance. can anyone guide me how to to this. the size of teh log file run in GB so i will need to take a daily backup
I have installed a linux server in my office to run 16 machines. Its main use will be a internal mail server but will be also running websites.
I have installed Ubuntu 9.10 server x64 and have got apache running.
I am looking for the simplest more robust solution for smtp, pop3 and imap. I have only ever used qmail before and found it a pain to configure and its getting old so I though I should probably try something new. I have not much experience with running pop3 or imap on linux so would love a suggestion on that.
I have an account in university on Linux machine with 10TB of free space accessible via SFTP. I would like to backup my Windows 7 x64 laptop to university. Currently I am using rsync+cygwin, but backup is pretty slow (without shadow copy) and I hate console window appearing every day on my screen when I login.
So I am looking for something like Windows Backup but with support for SFTP. Combination of tools will work too.
i setup the following script ro run each night (12am) as a cron on the main server:
if mount | grep -q '/home'; then rsync -ranv --delete /home/ 18.104.22.168:/home;
the main server is running on a dell poweredge 2600. The script rsyncs to a virtualised duplicate running on a hp dl380. when i set this script up and begn running data started going missing from the main server. if new files had been created the files by staff they would go missing, if data was added to existing files prior to activating the script the new changes made to that file would be lost.i just cant understand why this happened. as soon as i turned the script off after a few days it was all back to normal but the data that had gone missing had gone missing.
i just wanted to know if this could be a disk read/write issue, was the script running too soon not allowing data to be written before it can be backed up? could it bee memory? i just dont know. another developed occured after a few days of all this happening, which was one of our hard disk from the main server started mis behaving and flashing amber (attention).
We have a production web site running apache 2.2.3 across several web servers. we also have a major problem with SPAM comments right now. our method of identifying valid IPs (whether by external clients/customers, or internal personnel) vs SPAM'ers is not ideal - its prone to erroneously labeling legit IP's as targets to be blacklisted.
What we need is.. a way to see how much distinct request traffic is coming from any given IP address to the site in real time (or very near realtime). Essentially we want to see in some graphic/chart way requests per sec to apache / per ip sorted by requests per sec.Would nTop do this? I've only used this in a limited form at a branch office, not on a production web server.
I just installed RH5.4 on a sun server and I need to connect to it via SSH from a Windows server using a serial connection. Can anyone tell me how to configure SSH to access the server? What do I use for hostname and port #.Do I also have to install SSH on the linux server?