Ubuntu Servers :: Set Up A Backup Of The Windows 7 Pc?
May 16, 2010
I have been hassling with this for several days now. I have 64-bit Ubuntu Server 10.04 running on an Acer Aspire EasyStore H340. I have windows 7 running on a 64-bit desktop pc and on a laptop. I mainly wanted to use the Ubuntu server for a file server, so I installed Samba and created three shares. These do show up in Windows explorer, and I can read and write to them. My windows applications seem to be able to see the shares and open & save files.
My next step was to try and set up a backup of the Windows 7 pc to the Ubuntu server. Windows integrated backup sees the shares and sub-directories within the shares, and the initial part of the backup seems to run OK, but when it tries to save an image of the 'C:" drive it works for a long time and then ends with an error (cannot complete backup).
So, I looked for some free backup programs to try, but these do not allow me to select the shares as a destination (invalid destination). The dialogue sees the drives I have mapped the shares to in Windows, but does not show any sub-directories, and selecting the mapped drive letter does not take as a destination. If I try to browse down through "Network" in the destination dialogue, it selects "Network," but does not expand it or accept it as a destination.
So, I partitioned, formatted as ext3, and mounted my 2nd 1TB SATA drive on the server, and mounted it as "storage." I set this up as share in Samba and gave everyone read-write access, but still no luck selecting it for a backup destination. After some Googling, I downloaded and installed Ext2Fsd-0.48 (a windows 'driver' for Ext2/Ext3). This installed correctly, but when I open the program, neither "Network," the shares, or the mapped drives show up anywhere.
View 3 Replies
ADVERTISEMENT
Feb 17, 2010
I am mainly a Windows admin, but I do *NIX administration from time to time, for now I need to use an open source solution for backup windows environment mainly, I spent last days playing with bacula and backupPC, and I then chose backupPC, I built a solution with it seems working not bad, but before i go on deeper, I thought asking geeks here better, my main experience with back was with Vertias/Symantec Backup Exec, what do you recommend as most similar backup solution in Ubuntu offer a close level (I don't backup to tapes I back to hard disks), also a gui is preferred, while backupPC do a nice work and i handled its client config file (machine_name.pl), but I still do mistakes sometimes and troubleshooting is annoying, I have to backup files from users machines some of these files are running (like PST files), and I could need to backup a database or something from time to time. so whats your opinion all?
View 2 Replies
View Related
Jan 19, 2010
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
Code:
-- MySQL Administrator dump 1.4
--
-- ------------------------------------------------------
-- Server version`
[code]....
It seems that MySQL can open and write to the file fine, it just can't dump
View 3 Replies
View Related
May 10, 2010
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
View 7 Replies
View Related
Jan 19, 2010
I currently have a group of 3 servers connected to a local network. One is a web server, one is a mysql server, the other used for a specific function on my site (calculation of soccer matches!).
Anyway, I have been working on the site a lot lately but it is tedious connecting my USB hard drive to each computer and copying the files. This means I am not backing up as often as I should...
I have a laptop connected to this same network that I use for development so I can SSH into to the computers, is there any software for ubuntu that can take backups of files that I choose on multiple computers? I know I could rsync but is there something with more or an GUI?
Then I can just every 2 days move the most recent backup from my laptop to the USB drive. Then I will have the backup stored in 2 places if things go kaboom somewhere.
View 9 Replies
View Related
Oct 5, 2010
Is it possible in dd to use it for the output file to be stored at some remote location.
I do not have free space on the LVM partition whose backup I want to have via dd.
View 3 Replies
View Related
Aug 9, 2011
I have a file server running in an office that's mostly used for file sharing and a scanner saves pdf files to the server. I'm running the latest LTS ubuntu server edition and I really only have ssh installed and samba. My question is that I've done so much to the server as far as premissions and configuration and I'd like to make a clone of this to another computer and not sure how I would do this?
I'm not sure if clonezilla or something like this can perform this task? I basically just have a very old computer and now I have another very old computer that I want to make into a spare just incase something happens to the original. Any recommedations on how I would accomplish this?
View 3 Replies
View Related
May 3, 2010
I am running 8.04 Ubuntu server. Unfortunately a couple of days ago I thought I should upgrade as desktop upgrades usually go without a hitch and are very easy. I forgot that my server is live with a few websites and a radius server set up just the way I need (took painfully long time to figure out). Needless to say the upgrade caused many config file changes and many things stopped working. I panicked since this is a live server so I went straight to the backups to recover my system. I booted from a live CD and copied the entire system overtop the new one.
Everything that needed to work works, however now I get this message in my mail about every 10-20mins:
Subject: Cron <root@IMwebserver> [ -x /usr/lib/php5/maxlifetime ] && [ -d /var/lib/php5 ] && find /var/lib/php5/ -type f -cmin +$(/usr/lib/php5/maxlifetime) -print0 | xargs -r -0 rm
Content-Type: text/plain; charset=ANSI_X3.4-1968
X-Cron-Env: <SHELL=/bin/sh>
X-Cron-Env: <HOME=/root>
X-Cron-Env: <PATH=/usr/bin:/bin>
X-Cron-Env: <LOGNAME=root>
xargs: xargs.c:443: main: Assertion `bc_ctl.arg_max <= (131072-204' failed.
Aborted
Googling it, I found that its a problem with findutils. I tried to reinstall findutils with no luck.
My backup script looks like so:
@daily /usr/bin/rdiff-backup --exclude /dev --exclude /tmp --exclude /var/run/cups/cups.sock --exclude /var/log --exclude /mnt --exclude /media --exclude /proc --exclude /sys --exclude /var/cache/apt / /media/removable/BACKUP/rdiff/
How can I fix my system so the above e-mail no longer occurs?
View 1 Replies
View Related
Aug 29, 2010
I'm looking for way to automatically backup a few machines to my server. Does anyone know a good guide to set this up? I want it to pull the files from the machines at a certain time every week.
View 6 Replies
View Related
Sep 13, 2010
i want to run a postfix server as a backup mx, but anybody knows how can i collect the fist server mails with this one? this is multipop action but how can i do it with dovecot or any other pop collector?
View 1 Replies
View Related
Sep 15, 2010
I have a Linux host acting as an ISCSI server for a Windows box. I want to keep an off site backup, so I figure rsync will keep the ISCSI server synced with an offsite Linux host. I understand that Rsync does block level incremental transfer to conserve bandwidth ok, awesome.The trick is, that I also want an archival copy kept. Say I want to go back to a revision of a file from 10 days ago, I need to be able to do that.
I was planning on using Backup Exec, since we currently have a licensed copy. Throw the archives from Backup Exec onto the ISCSI server as well, and have it keep a rotating 30 day backup, or something like that. The issue I see here is that this will be creating a deleting files as it does its daily backup rotation. I'm guessing RSYNC will see these as new files, and likely retransmit everything on a daily basis. The question then becomes, is this assumption correct, or will it still know to do a block level incremental transfer even when file names and such are changing?
View 5 Replies
View Related
Oct 21, 2010
I have two shares in total and there are also two external hard drives. The server is used by two different organisations that are not supposed to have access to the data of the other one(at least not as normal users).he script I need should run in the background of the server and when a drive is plugged in, it should check, which organization the drive belongs to, and depending on who the drive belongs to, backup the respective share.When the drive belong to neither, it should just do nothing.Unfortunately, I have no clue about scripting and so this makes writing a script like that, at least for me, impossible.So I wanted to know if somebody could name some good websites for learning to write such a script or give tips.
View 9 Replies
View Related
Feb 4, 2011
I support a small business which has an Ubuntu server running as a file server. The server is running Ubuntu 10.4. There is one hard drive which is mounted as /media/hdd. Each night this is backed up to an external USB hard drive mounted as /media/backup. The backup is carried out using the command:
Code:
rsync -av /media/hdd/ /media/backup/
Is there a way to encrypt this back-up so that if the USB hard drive is plugged into another machine it cannot be read?
View 5 Replies
View Related
Mar 7, 2011
I want to make a daily backup of my websites from ubuntu server over ftp to another server I own. Backup schedule and process works, the problem is backup restore. Winrar says: The file is corrupt, 7-zip crashes.
Backup archive looks ok (the same size as original folder) and you can also extract it ignoring the error by winrar. But the extracted folder only contains one or two subfolder and one file(usually image) and that's all.
If I try to restore from Webmin it doesn't report any error, and it looks like restore worked. But restored files are nowhere.
View 5 Replies
View Related
Jan 3, 2010
How can I work your backup to Linux system without Windows
View 8 Replies
View Related
Mar 1, 2010
I have a personal ubuntu server that provides apache, glassfish, firewall, routing, email, CVS, MySQL, etc.... This server has been running for a while with two hard drives configured into a RAID 1 array. The array has two partitions, one for swap and one for the data. I currently back up the data with a removable hard drive. I use dd and create an image of one drive and the MBRs (partition tables) of each drive.In a disaster situation I can use this data to recreate one drive and then re mirror it to the second, or just boot the back up.I like this solution because I can easily recover from bare metal, and the backup is transparent. I can browser it if needed since its an uncompressed image of the drive. The one drawback is that I need to reboot the system with a linux CD to do the backup.
My hard drive space is almost at capacity. So what I want to do is add a third drive to the array and migrate it to RAID 5. However this will cause my current backup method to no longer work. How can I back up this RAID 5 array. I need to back up the entire system, and not just the data. I have made many tweaks to the system over the years that it has been running that I can't lose if a restore is needed. I have seen a large thread here that people have been using tar. My concern with tar is how do you use a tar archive to restore a system to a new array. Im assuming that you would need to setup the array and then just restore the archive? Also, i don't have much faith in using tar on a running system. Doesn't this open yourself up to corrupted backups? My second idea is using rsync. While I consider myself experienced in linux from 10 years of personal and professions use, I have not had much experience with this utility. Would rsync provide a more reliable way to backup a running system that would enable a bare metal restore later? I once read something about people using rsync with hard links to create a backup that could store many incremental backups.My main concern with both rsync and tar is not being able to restore the OS to the state that it was in at the time of the backup.
View 4 Replies
View Related
Aug 8, 2010
Does ubuntu server automatically backup files? How can i get to them?
View 1 Replies
View Related
Aug 10, 2010
I'm going to make a nightly backup copy from one server to another, using rsync. If I have a sufficiently large file, say 4+ GB or so, I'm not interested in copying the whole file if only a small change has been made. Can rsync detect small changes on block level and backup only those if needed?
View 2 Replies
View Related
Oct 12, 2010
I currently have an Ubuntu 10.04 Server with 10 2TB hard drives (Hot Swappable). I discovered that having a software raid over 16TB is not supported, so I split the drives into 2 sections and have 2 Software Raid arrays storing my movies, audio, pictures, and other software. The total current usage is around 7TB.
Since backing the files up to DVDs or even BlueRay is laughable, I am going to backup the system to 2TB hard drives probably 4 of them, the problem becomes that I can only hook one backup drive at a time into the system using a hot swap tray. Now I know that I can do this manually by copying the files one at a time to the drive until it is full, switching the drive out and repeating this, but I am hoping for an automated solution, Start backup, plug first drive in, system fills up drive, swap and repeat. Also it would be nice if the system remembered what had already been backed up so when I add files to the system, I only need to attach the last drive and not start the process over.
View 3 Replies
View Related
Mar 3, 2011
I'm looking for a way to do a bare metal backup of our server using a tool such as ghost or clonezilla. The limitation is that / is on an mdadm raid 5. The only relevant info I could find on clonezilla's site was:
# Software RAID/fake RAID is not supported by default. It's can be done manually only.
View 6 Replies
View Related
Mar 21, 2009
I'm going to be launching my website soon, and I found a company to host it on one of their dedicated servers. I think I'm going to go with fedora as to OS, and well my problem is I'm having trouble finding a company to backup up my files, that both supports fedora and well is reasonably priced.
View 2 Replies
View Related
Oct 7, 2010
I have several machines that I'd like to backup various folders through the linux box onto DVD-RW media.I want to keep log files of what was written and when to dvd on one machine and have it automagically assign a unique serial number that I can print on the dvd in case I need to recover.I'd like a user friendly UI that I can point and click to schedule the backup and it's type.Is there a good Fedora backup application (read easy to use/understand and configure) I can use to backup machines across a network and across multiple dvd's (if needed)? The host machine is a F11 box.These are a mix of Win server 2008, win 7, win XP and several Fedora boxes.Speaking of dvd media is this a good idea and how many erase/write cycles are they good for?
View 2 Replies
View Related
Aug 8, 2011
I need to install a backup server in my work environment.We have a Windows 2008 server and an old DELL PowerEdge 1750 server that has no OS on it yet.I would like to install Fedora on it and then backup the Windows Server data on the Fedora server using rsync or something else to do the backups.Do you think it's a good idea ? If not what would you use to backup the Windows server data, preferably on a linux system.
View 12 Replies
View Related
Jan 11, 2011
I got a Web Server running great. But I need to back it up. It doesn't for some reason detect either of my USB External HD's. How can I make a TAR or something and burn it to DVD and then be able to restore using that disk.
View 6 Replies
View Related
Jun 28, 2011
I am currently working on managing multiple linux servers in remote locations, servers particularly user for web hosting. I need to backup data to a backup server but rsync which i currently using doesnt helps is there any tool to backup every server with out modifying it bcos there are hundreds of servers so installing a tool in every server is time consuming process.
View 7 Replies
View Related
Feb 24, 2010
I administer several web hosting (combined with mail relays and other services) production servers under Debian GNU/Linux. I began giving these public services two years ago via three boxes: the first is a gateway which controls traffic via iptables (it's attached to a DSL modem) between a public subnet (the DMZ) and a local network which connects several workstations. In the DMZ subnet I maintain two Pentium-III era boxes, they've grown in services since I set them up. Actually, I think I should buy new ones, but, you know, I want to save money and lenghten its life.
So, they've grown in data hosted, but I've never implemented a resilent backup system. I've set up some rsync tasks sheduled via cron jobs to copy the entire UNIX file system in each of the DMZ boxes, but I'd like to be prepared before an unexpected "real" crash of some HDD, I mean, some problem that renders a disk unusable.
AFAIK, sysadmins sync entire HD backups which are capable of recovering a system via swapping the unusable unit with the backup unit. Maybe the best fashion is to implement a RAID mirroring the unit, I'm I right? So, keeping my systems as they are, I mean, capable of using 4 parallel ATA units, what would you do? Use dump, rsync or some other way to have an operational second unit with an exact copy in a bootable second drive, in order to quickly swap it if the main unit fails?
Comes to my mind to partition a second unit (so making it bootable) and backup daily via rsync only those parts of the unix file system hierarchy which are necessary to boot a system properly. What do you think about this workaround?
View 6 Replies
View Related
Jan 17, 2011
It says "This backup could take up 172 GB of hard drive space" and my C: "Used space" is at about 172 GB. So will the Ubuntu partition be backed up?
View 3 Replies
View Related
Jun 15, 2011
I had some problems with my hdd, I fixed most of the problems with "testdrive" or "testdisk" (cant remember) but I have some bad sections on my disk right where my important data is. I made a backup with "dd" from a live cd. I have the backup with a reading error or 5 so I guess most data is safe? I have the backup of the whole drive on an other hdd and now I need to get just a few folders out of it. I tried in windows but I can't read it. Poweriso can but not the privet folders like /var/ and that folder I need. I had a password on the account so I guess that is part of the problem? How to read the backup file in windows?
View 4 Replies
View Related
Jan 23, 2010
After a near miss with my 1.5TB, RAID5 file server, I have decided that I need to backup my data to an external hardrive periodically.I have been looking at rsync but the question I have is: Do I format the external hard drive in EXT3 (the sameas my fileserver) or NTFS?All my main machines are Windoze, but the file server is Ubuntu with a samba share.If my server ever went belly up, I would like to be able to access my data from the external hard drive. I guess if it's in EXT3 then windows would be clueless... I would either need to fix the server pronto or access it with a live CD or something.What would I lose if I used NTFS instead of EXT3? I think I would lose permissions and possibly ownerhsip information - are there any other issues?
View 3 Replies
View Related
May 11, 2010
After my 8.04 debacle here:I basically gave up on trying to write an install guide for Hardy and Amanda. Others had done it better and with how muddled I was by the end of getting a working server, I wasn't sure I could repeat it.Now in Lucid, I've managed to get another working amanda server going, I'm fairly sure I can repeat it and but I question:Installation with 'apt-get install amanda-server' uses the user backup in the group backup.Installation of the ZWC client uses the user amandabackup. while thankfully ZWC has included reg keys:
HKEY_LOCAL_MACHINESOFTWAREmandaWC1.0Install BackupUserHKEY_LOCAL_MACHINESOFTWAREmandaWC1.0Install RecoveryUserto change the user on the client, this is messy, requires regedit (which normally is locked out by group policy), editing the service in service.msc and creation of a new user account named backup in addition to the installer's default on each client.The actual question is:1) Is there a reason for the use of backup:backup? (rather than amandabackup:disk or amandabackup:backup)
2) Is there any sort of flag that could be set, script to change the user within the repo atinstall time.Or generally an easier way of doing this than compiling each time I do an install.
View 1 Replies
View Related