General :: Use The Second Drive For Backups, Both Manual And Using Back In Time \ Noticed That The Second Drive Was Formated In FAT32?
Nov 1, 2010
I have a second hard drive in my desktop and both the main dirve and second drive are 250GB. I use the second drive for backups, both manual and using back in time. The other day I noticed that the second drive was formated in FAT32. If I go to disk utility and look at the drive it says:Usage:FilesystemPartition Type:Linux (0x83)Type:FAT(32-bit version)Is this ok? I thought in Linux it should be Ext4. So far its been working fine for a while now but if I need to move my files and re format it to ext4 and move them back I would rather do it now when there is less data on the drive.
I loaded Ubuntu 8.4 on a data drive (second drive no OS) from a Windows XP-SP3 system. I MEANT to load it on the main XP OS drive. Bottom line I formatted a FAT-32 with Ubuntu 8.4. Can I (freeware hopefully) roll back the Ubuntu formatted drive to FAT-32 so I can recover my data?
I have a small IT consulting business and I am finding many of my clients couldn't afford Microsoft solutions and thus, going without... So a revelation hit me that I could offer Linux solutions, its just that I'm not a Linux guru... So after much research and installations of nearly every latest Linux distro, I decided on ClearOS as a good option for my clients that just require a good File/Print server, firewall and VPN solution... Everything was going fine until I got to the point where I was trying to get it to do backups to a USB drive... ClearOS does not come with the ability to do such, so from further research I found that I could install Webmin to handle that task and that it would not break ClearOS... Great, its just that its not working... It says its working, but its not.
Next I hit the "Save and Backup Now" button to test it out... all says it goes good... But when I check the device, there's nothing saved to it...
So I create another "Filesystem backup", I select the FlexShare directory that I want backed up and this time I select a local directory for the backup to go to... I hit the "Save and Backup Now" button to test it out. all says it goes good... But once again when I check the directory, there's nothing saved to it.
I have a small IT consulting business and I am finding many of my clients couldn't afford Microsoft solutions and thus, going without... So a revelation hit me that I could offer Linux solutions, its just that I'm not a Linux guru... So after much research and installations of nearly every latest Linux distro, I decided on ClearOS as a good option for my clients that just require a good File/Print server, firewall and VPN solution... Everything was going fine until I got to the point where I was trying to get it to do backups to a USB drive... ClearOS does not come with the ability to do such, so from further research I found that I could install Webmin to handle that task and that it would not break ClearOS... Great, its just that its not working... It says its working, but its not...
So here's what I got, ClearOS 5.1 with Webmin 1.5
Here's what I have done to try to make it work after a good installation...
Under "Hardware" and "Partitions on Local Disks", it shows the USB drive as Device B... So I create a partion called "/dev/sdb1" Great, I'm thinking...
So I go to "System" and under "Filesystem backup", I select the FlexShare directory that I want backed up and then I select "/dev/sdb1" as the backup to device...
Next I hit the "Save and Backup Now" button to test it out... all says it goes good... But when I check the device, there's nothing saved to it...
So I create another "Filesystem backup", I select the FlexShare directory that I want backed up and this time I select a local directory for the backup to go to... I hit the "Save and Backup Now" button to test it out... all says it goes good... But once again when I check the directory, there's nothing saved to it...
I have inherited a network server running Redhat Enterprise (4?) which uses an external USB drive for backups. These have been scheduled to run at midnight each night.I want to use 2 external drives to hold the backups (exchanging the usb drive each day). My question is: Is there anything I need to do to a new USB drive before exchanging the drives (eg. formatting etc) or can I simply just plug the new one in and let it run?I apologise for the very basic nature of this question but I have no clue about Linux.
I have an external harddrive which is fat32 (which was filled when I was on a windows system) with a lot of files on (> 200gb) which have mixed case filenames. I wanted to write a script to rename them all to lowercase however when I did I got an error saying that the filenames are the same (which I guess is true as FAT is case-insensitive). My issue is that I'm mounting this from a linux box now so it would make it a lot easier if things were lower case! I've already run a script to replace all spaces with underscores.
I know I could do this by first going through and renaming everything with a prefix and then renaming it back again (ie, rename File.Jpg to xFile.Jpg and then a second rename to strip that x off and rename lowercase to file.jpg) however I find this approach a bit messy and would prefer to do it all in one pass. Other than that, could I change the drive to ext4 without losing any files on it? The drive has a hell of a lot of stuff on there which is an archive of many years of files - I'd be absolutely gutted if I lost everything.
I need to get a backup copy of a huge directory on one of our RHEL servers. Rather than hook the external USB drive up to my PC and manually copy it all across the network, can I just plug it in to one of my USB ports on the server and rsync it?I know how to do the rsync, I just don't know how to get the USB drive to show up when I do a df -h and how to properly remove it before unplugging it from the server. this is a live server, so I can't go playing around and possibly mess something up.
one would have to exclude certain folders / directories but would the backup be possible if the system is up and running in its native "live" state ? Which directories could be excluded ? Does swap need to be turned off ? I would like to make incremental backups on a separate partition of the same hard drive. I will endeavour to backup the MBR/ Partition table using dd.
I am working from a laptop where all my work is stored on a 80GB drive. I am now also an owner of an external 250GB USB hard drive, formatted with FAT32. I want to keep it FAT32, so that I can offer some of my files to people that run Mac OS or Windows and I don't want to have them install ext3 for windows and what not.I am in need of a strategy which will allow me to keep a mirror of my laptop drive on my new external drive, i.e. no history / versioning required. However, I do care about file permissions. The files don't have to be stored as-is, they can be stored within a large (80GB?) tar file, that is fine - it would be easier for me to coerce people to open a .tar file than to install an ext3 driver for their OS, I suppose. I don't think I can keep file permissions otherwise, can I?
I have previously used a self-written sh script that used rsync to keep an up-to-date copy of my laptop filesystem on a USB flash drive, but in that case I had the flash drive formatted with ext3, so no problem with file permissions there. This time, it's trickier.
I am using Debian Lenny 5.0.3 with a stock 2.6.26-2-amd64 kernel, and have a ~/Documents/HTMLS directory of 273.2MB (21590 files, 1063 sub-folders) which according to konqueror shows 1130 items - 572 files (60.8MB Total) - 558 folders, these were created by 'Save Page As' in Iceweasel 3.5.11. I am trying to copy the contents of this directory to a similar directory on a 2GB USB thumbdrive /dev/sdb1 which was partitioned and formatted as fat32 by Qparted. Problem is that the copying ceases after about 6 files transfer. I found that (as I am sure you know) the named.html files come with a matching named_files folder, and often there are what appear to be invalid characters such as '*','?',and ':' in the filenames in said named_files folders.
After exhaustive googling [I know we all say that] I found an instruction: Code: pax -rw -s '/[*?:]/_/gp' stuff /fat32/partition that changes the name of the files, replacing said characters with '_', but whilst I can get pax from my repos, I really don't want to 'archive' the files - since I understand that pax was created to bridge a war between tar and cpio - because I want to be able to read the html files on an old (not connected to the Internet) WinXP tablet. So, I believe that I need to create a script, that scans all the filenames, greps and seds to replace said 'unacceptable' characters. I am assuming that Firefox on the Tablet PC will be able to open the htmls if I can get them onto the thumb drive. Are there any other known characters in filenames that M$ file systems can't handle?
Dropbox will not start properly because my Lucid installation is on a SS HD (/dev/sdc) but my data, including my Dropbox folder is on an internal NTFS-formatted HD (/dev/sda), and I also have another internal HD for backups (/dev/sdb).
For some reason I can get the backups HD to auto-mount on startup, but not the data HD. My fstab file looks like this:
Just installed opensuse 11.3 Kdeversion on my laptop. Before installing it on live mode i had a problem of accessing my other drives (NTFS, FAT32 and EXT4) which said HAL system policy...etc mounting error. I could access all drives with root privilege. I thought problem will be solver once i install opensuse on my system. How ever i was really disappointed after seeing the same problem post install. Googled around for the solution and got this link
After this the problem got worse now i am not able to see any of the drives in the side panel. Gone through many forum and posts all discuss about external USB HDD.
i needed to change my external hard drive's file system from ext3 to fat32, to use it in windows, which i did the simple way: i shrunk the ext3 partition, made a fat32 partition, copied the files over, removed the ext3 and made the fat32 bigger. unfortunately, while gparted was making the partition larger, my computer shut down. i lost all my files and the partition messed up immediately. i made a new fat32 partition, after deleting the old one, but noticed that gparted was showing 100 gigs already in use (???). so now i have a 300 gb hard drive with only 200 gb i can use; i ran df to make sure gparted wasn't messing up, but indeed it shows the partition as being only 200 gigs in size. i haven't tried making any other kind of partition yet, such as ext3, for fear of losing my files again, and because it wouldn't be permanent anyway, because i need those files in windows and stupid microsoft won't make their OS ext3 compatible.
I need to know when I have bought a notebook. I know I have formatted disk myself and partition created and Ubuntu installed.Is there any way I can get info when I bought a notebook? Like time of disk formatting, partitions created, Ubuntu installed
I am using Asus eee PC model 1001P and I am running ubuntu netbook remix on it. I have a 160 GB hard drive. When I was installing ubuntu I assigned only 50 GB for it, thinking that I will use the remaining partition for windows. But right now after using UNR, I don't want to use windows at all. Is it possible that I can assign the remaining 100 GB unassigned partition back to Ubuntu without reinstalling the whole operating system again?
formatted a 64g usb memory stick with gparted use fat32 file system and recognized full 64gig drive plug in and is on /media/usb this is what i got on my fstab UUID=24AF-1E67 /media/The-Hive defaults,nosuid,nodev,locale=en_US.UTF-8 0 0 i want all user access to this memory stick to write delete and execute. am I going nuts i cant make it to do that..
I've picked up an HP Simplesave external drive. It comes with some fancy software that is of no use to me because I don't use Windows. Like many current consumer-targeted backup drives, the backup software is actually contained on the drive itself. I'd like to save the drive's initial state so that I can restore it if I decide to sell it.
The backup box itself is somewhat customized: in addition to the hard drive device, it presents a CDROM-like device on /dev/sr0. I gather that the purpose of this cdrom device is to bootstrap via Windows autoplay the backup application which lives on the disk itself. I wouldn't suppose any guarantees about how it does this, so it seems important to preserve the exact state of the disk.
The drive is formatted with a single 500GB NTFS partition. My initial thought was to use dd to dump the disk (/dev/sdb) itself, but this proved impractical, as the resulting file was not sparse. This seemed to be because the NTFS empty space is not filled with zeroes, but with a repeating series of 16 bytes.
I tried gzipping the output of dd. This reduced to the file to a manageable size — the first 18GB was compressed to 81MB, versus 47MB to tarball the contents of the mounted filesystem — but it was very slow on my admittedly somewhat derelict Pentium M processor. The time to do that first 18GB was about 30 minutes.
I have a mini ITX computer that runs a software that I have been trying to clone to another hard drive as back up. When I do clone it, it is the exact same copy down to every byte. When I try to boot from the cloned hard drive, it starts booting, then displays some text:
Then is says some of the files are not found including a encryption signer key not found.
I do no know the next step I need to take. Is this becasue it is running off a different hard drive? I do not understand a bzImage too well, but does it have anything to do with that file?
My media library isn't huge, but it isn't tiny (~50 GB). Every month or so, I just manually copy ~/Music, ~/Pictures, and ~/Videos to my EHD, and delete the old backup. But this is far from ideal. It's pretty slow, for one thing (~50 GB all together). It also isn't versioned, so if I ever want to go back multiple versions, I'm out of luck.
Is there any simple, stable, incremental way to do this? I'm open to using traditional version control systems like Git for it, although I haven't used them before for anything other than code. Command-line is fine (especially if it's scriptable). I only need to back up these 3 folders--anything that's not media is stored in my Dropbox.
I have an OpenBSD and a FreeBSD system and a mac. I also have a Linux server. What i would like to do is back up all these systems to an external hard-drive using rsync when the external usb disk is connected to my linux box.If i format the external usb disk with cfdisk and the create a non-bootable ext3 file system on this external disk and create and put all the necessary public keys on the Linux box then from the BSD's or the mac issue the command:
Will this back up the entire systems so that they can be restored in the event of an emergency? I should store each OS just in a separate disk file of the external usb drive each time right? Because i would rather not have to format the external usb drive for each different OS. Would this work? and would the restoration command for these BSD's be:
I just need to know the basics. I'm sure given that i'll be able to automate the process. I don't want to clone the disks for forensics. I just want to have a way of restoring to a clean OS. This is the most basic question:All the howto's never mention whether or not you have to have an rsync server running on the machine your backing up to. So do you just push or pull from one end of the connection only or do you have to have a client at one end and a server at the other, as is traditional?
I have a USB drive on which I want to install Arch Linux (using the installer, not unetbootin or something similar, as I want the drive to be persistent.) The computer from which I want to boot this USB supports booting from a USB floppy, not a normal drive. Is there any way for me to make a USB floppy on another drive and use that to boot the normal USB drive?