I have 2 external drives that I'm using for movies and they're both formatted to fat32. The problem I'm experiencing now is that some of the high quality mkv files are not able to be moved to the drive because of fat32's limitations.
I'd like to convert to ext3 (or 4) and I wanted to check to see if I can do it without having to reformat and without loss of data.
is there a way of sharing an ext3/ext4 formatted partition on an external USB drive between different users (uids) on different Linux machines without creating a group for this purpose, setting the group ownership of the partition to this group and adding each respective user to the group on every machine?This would mean that I need to have root privileges on every machine... which I may not have in some cases.I'm using the partition to store the code I'm developing on Linux and I would like the option to be safe... if possible.I could use a vfat partition but then I have no control of the rw rights + I cannot develop directly in the dir: I would always have to tar.gz the directory, extract, work, tar.gz, copy to the external drive.
I am wondering if any of you technical guys would be willing to format my Western Digital external USB 1.5 TB Hard Drive to Linux EXT3. I am naturally happy to pay for your time and trouble and for postage. The WD drive is for storing video footage and will be connected to my Humax Freesat HD Digital TV Box(not a computer), and the Humax Box will only record high deffination programmes in EXT3 format. I've tried to do the job myself with my PC, but have failed to change my system to format in Linux.
I have a Western Digital My Book World Edition external harddisk with blue rings. I filled it up and now want to delete the data and start over. I'm set in my ways and have been accustom to reformatting harddisks periodically (sector maintainance, etc.) It's worked for me as I've luckily have not had a disk crash in 25 years.My webapp is not helping me with the reformat and neither is Western Digital tech support. I've heard that it was factory formatted with something called Linux ext3. Does this make sense? Has anyone had any experience with reformatting external harddisks being used as a NAS (home use).
I am building a home server that will host a multitude of files; from mp3s to ebooks to FEA software and files. I don't know if RAID is the right thing for me. This server will have all the files that I have accumulated over the years and if the drive fails than I will be S.O.L. I have seen discussions where someone has RAID 1 setup but they don't have their drives internally (to the case), they bought 2 separate external hard drives with eSata to minimize an electrical failure to the drives. (I guess this is a good idea)I have also read about having one drive then using a second to rsync data every week. I planned on purchasing 2 enterprise hard drives of 500 MB to 1 GB but I don't have any experience with how I should handle my data
I suspect this is not new but I just can't find where it was treated. Maybe someone can give me a good lead.I just want to prevent certain users from accessing CD/DVD drives and all external drives. They should be able to mount their home directories and move around within the OS but they shouldn't be able to move data away from the PC. Any Clues?
Sometimes I get a problem with the basic "umount" command. I get an error message telling me that for one reason or another, the device couldn't be unmounted. Usually, it says the device is busy, when I can't see how it possibly can be. When this happens I'll use "umount -l" or "umount -f" or sometimes "eject" but I'm still not happy, because at the back of my mind I'm worried about damaging the integrity of the device's filesystem. What's the CORRECT way to deal with this problem?
My Motherboard has 4 SATA ports on it, Is there a way short of buying an expensive RAID card to add more SATA drives and do a software raid still? What about getting an external 8-bay eSATA enclosure and putting drives in it? Will the OS see this and software raid? (linux)
I've recently brought a Western Digital Elements 2TB external hard disk and have been planning to encrypt it for use as a backup drive. However, it seems that these 2TB disks use the new 4K sector sizes and thus need to be handled more carefully than the older 512K ones.
After spending a week looking on Google, I have to admit I'm pretty confused and hope somebody here might be able to verify my conclusions
The drive reports that it's a 512-sector drive which is probably false. Using fdisk -uc, the original partition starts at sector 2048 so I assume that is a valid sector also to start a dm-crypt partition overwriting the previous one?
I've also read that every layer that is added to these drives must support the 4k layer. That means both dm-crypt and the ext3 filesystem I intend to put it on have to do so also.
Looking through the cryptsetup document, it states under the option "--align-payload" the following:
"Align payload at a boundary of value 512-byte sectors. This option is relevant for luksFormat. If your block device lives on a RAID, it is useful to align the filesystem at full stripe boundaries so it can take advantage of the RAIDs geometry. See for instance the sunit and swidth options in the mkfs.xfs manual page. By default, the payload is aligned at an 8 sector (4096 byte) boundary."
The fact that the payload is aligned at 4096 seems to indicate to me that it should be fine using default settings. Does everybody agree with this? Or do I need to take special measures due to the dm-crypt headers?
When I later finish up the dm-crypt layer, then I need to put ext3 on it. I understand adding -b 4096 to the mkfs.ext3 command will resolve that. Is that also correct and will it work well in combination with the dm-crypt layer?
I have a FAT32 external USB hard drive with a bunch of stuff I want to copy onto a RHEL server. Is it as simple as it is on a Mac or PC where I just plug it in and it will show up, then I can copy all the files off of it?If it is, how do I safely remove the drive after I'm done with it?
It seems to be (we have experienced) that 10.10 can't handle having an older ext3 drive mounted to a folder on the 1010 ext4 drive. The files get "corrupted", but afterwards, putting the drive back into the original older 9.04 system, the files are perfectly fine again.
I am hosting a few customer servers now, all of which are virtual machines running on a CENTOS 5.x host. Each Centos host has a couple of extra drives. When I formatted them ext3 they automatically had a schedule of a full forced fsck after 6 months. Do I really need to do that check regularly? It results in a fairly large outage since my disks are each 1TB and there are up to three extra drives on each server. I try to reboot these servers every 6 months but this part adds a large amount of time to a routine reboot.
I've lost my admin password on my current Windows OS and would like to install Linux Ubuntu or a similar user-friendly distro of Linux alongside, see how that goes and possibly reformat my PC with Linux as I was told it would convert NTFS formatted drives to ext3, not delete them.
A friend of mine and I bought an external hd (WD Elements 2TB) and formatted it in ext3 as root. Now we want to use this hard disk in more than one systems with different usernames. So we did "chmod -R 777 /media/ VolumeLabel" in all the systems.But we want the hd to pass around. So its a little bit inconvenient to do chmod all the time. So what should we do to make the access for the hd universal in all the systems that we plug it in?
I have a Westel 1 TB external hard drive and when attempting to partition it, the partitioner never finishes refreshing the device. I've tried gparted live, opensuse, & ubuntu; and none of them finish refreshing the device.
Another issue is, I have 200 gig of music & movies that can't be wiped out because I got no room elsewhere to move them to. I would like to set a partition of 750 gig for the ext 3 and leave the rest for ntfs.
I have a external HDD with eSATA and USB connectors available. I want to use this HDD to store my backups. The HDD should be encrypted (my main system is as well).
So here is what I did so far: 1) I used the following code to create the encrypted LUKS partition with EXT3 Filesystem: Code: cryptsetup -c aes-xts-plain -s 512 luksFormat /dev/sdb1 cryptsetup luksOpen /dev/sdb1 luks mkfs.ext3 /dev/mapper/luks The system always hang when I executed the "mkfs.ext3..." command, so I switched the HDD from eSATA to USB and then it worked fine.
2) When I switched on the ext. HDD the first time, the drive was recognized automatically and Nautilus asked for the password. I typed it in as checked the checkbox to remember the password in the future. For the backup I use a nice script that I found in another forum, where I can define a mountpoint and then the script will check for previous backups and only make a incremental backup based of the latest version. The script also mounts the drive automatically. In order to always have the same mountpoint, I want to make an entry in the /etc/fstab using the UUID of the ext. HDD.
Whatever I tried, it doesn't work. What am I doing wrong? Here is my current /etc/fstab Code: # <file system> <mount point> <type> <options> <dump> <pass> proc /proc proc defaults 0 0 # / was on /dev/mapper/ubuntu-root during installation UUID=2ea47421-73ce-4c66-9606-8a1db81ae640 / ext3 relatime,errors=remount-ro 0 1 # /boot was on /dev/sda1 during installation UUID=dbdeb793-1d4e-43ea-8986-7b37fdbc9674 /boot ext3 relatime 0 2 # /home was on /dev/mapper/ubuntu-home during installation UUID=42702091-83e6-43eb-aad1-108f43eedf9d /home ext3 relatime 0 2 # swap was on /dev/mapper/ubuntu-swap during installation UUID=e225bcf9-908b-4226-a963-6b02ee658df1 none swap sw 0 0 /dev/scd0 /media/cdrom0 udf,iso9660 user,noauto,exec,utf8 0 0 # Eintrag wegen iPhone none /proc/bus/usb usbfs devgid=125,devmode=666,nodev,nosuid,noexec 0 0 # external HDD UUID=913977f7-8fa6-416f-af79-b5f913b68f53 /media/backup-hdd ext3 noauto,users 0 0 I made the "none /proc/bus/usb..." entry because it was recommended to ensure correct behaviour of the iPhone. Not sure if I need it though.
I created the mountpoint with this command: Code: sudo mkdir /media/backup-hdd Now it seems the mountpoints owner is not root - strange right? Code: 2 4 drwxr-xr-x 3 michael michael 4096 2010-01-15 02:45 backup-hdd How should I mount this drive correctly? It will be automounted as every USB device, but that should not be the case. I want the script to mount and unmount the drive.
I'm trying to format a 500 GB external drive with gparted in ubuntu 10.10 (I searched & didn't see this issue in the forum). I set up and formatted two partitions, one for fat32, and the other with ext3, which appears to format ok, but I can't use it. Both partitions show up and appear to mount, but the ext3 partition won't accept activity (make new folder, copy in files), while the fat32 partition works fine. Both partitions show up ok when I query in terminal "sudo fdisk -l"
I have an Iomega External Hard Drive 1TB. Problem: Unable to write to Ext3 Partition. How I got here: Started off with going into windows and shrinking it's current NTFS partition down by 50GB. Then used an Ubuntu LiveCD to gain access to GParted and with that 50GB free I formatted it with an Ext3 system. It does this no problems. I then can't write anything to this partition? I've tried doing "chmod 777 /dev/sdb2" but it says Operation not Permitted.
EDIT: Need to be able to use it on YDL 6.2 on my PS3, YDL doesn't have NTFS write support and I want to transfer files larger than 4GB rendering Fat32 useless. No I'm not able to re-size or compress this large file.
After a near miss with my 1.5TB, RAID5 file server, I have decided that I need to backup my data to an external hardrive periodically.I have been looking at rsync but the question I have is: Do I format the external hard drive in EXT3 (the sameas my fileserver) or NTFS?All my main machines are Windoze, but the file server is Ubuntu with a samba share.If my server ever went belly up, I would like to be able to access my data from the external hard drive. I guess if it's in EXT3 then windows would be clueless... I would either need to fix the server pronto or access it with a live CD or something.What would I lose if I used NTFS instead of EXT3? I think I would lose permissions and possibly ownerhsip information - are there any other issues?
I Tarred and GZipped most of the data on one 1Tb partition and stuck the archives on a second 1Tb partition on a separate disk. I then proceeded to format the first partition with NTFS (from Linux.) The only problem is that I completely forgot that I had a CD drive and formatted sdc1 instead of sdd1! I began doing a full NTFS format and after a minute or two I cancelled it and decided to do a quick format. I then realized my mistake. I managed to find a copy of the superblock and began trying to recover the disk. fsck -t ext3 recognized the partition as NTFS but I luckily didn't have fsck.ntfs installed so it didn't touch it. I managed to get it working with fsck.ext3 (with -b,-B and -y) fsck.ext3 didn't mind that it was an NTFS partition.
Roughly how long will this take? It's running from Knoppix within a virtual machine to a USB hard drive which is 100% full. Days? Being that for a few minutes I attempted a full format am I going to end up with a bunch of corrupted archives? If I do end up with file corruption can anyone recommend a way of recovering the data / sorting it out? Is it likely to be just a few old files that are corrupt (It's my understanding that filesystems like to keep files in the same area on the disk to minimize the amount of head travel.) This might just be wishful thinking but as the filesystem fills up will ext3 put the newer files towards the end of the disk? If so then I'm hoping that a full NTFS format starts at the beginning of the disk.
i needed to change my external hard drive's file system from ext3 to fat32, to use it in windows, which i did the simple way: i shrunk the ext3 partition, made a fat32 partition, copied the files over, removed the ext3 and made the fat32 bigger. unfortunately, while gparted was making the partition larger, my computer shut down. i lost all my files and the partition messed up immediately. i made a new fat32 partition, after deleting the old one, but noticed that gparted was showing 100 gigs already in use (???). so now i have a 300 gb hard drive with only 200 gb i can use; i ran df to make sure gparted wasn't messing up, but indeed it shows the partition as being only 200 gigs in size. i haven't tried making any other kind of partition yet, such as ext3, for fear of losing my files again, and because it wouldn't be permanent anyway, because i need those files in windows and stupid microsoft won't make their OS ext3 compatible.
To force myself to use CLI all the time, I've installed Ubuntu Server edition on my work laptop. However, I still want to be able to view PDF ( rendered as normal PDF, not converted to text ). Basically I want a totally CLI environment, except for PDF viewer.
I want to generate HTML files in batch to show some gvim colorschemes in action. I took my current solution from the script at but this will open open a gvim window for each colorscheme I want to process. So far, the only way I have found to avoid the annoying new windows that pop up every second is to start a VNCserver and set the DISPLAY environment variable to that of the VNCserver so that all gvim windows are sent to the display within the VNC session.However, I would like to know if there is a way I can avoid the whole VNC setup and just run a headless gvim instance that does the conversion and exits, with no windows ever being actually displayed.
I can't use my shares on external drives. I've tried to work with this and I really like Fedora but enough is enough. enough having to recover from multiple crashes attempting to get Nvidia drivers installed but I've given up on this. I haven't found a solution after HOURS and HOURS of searching. I was willing to wait for 13 to have 3d support on my Nvidia card but if I can't find a solution to this one
I have recently installed Ubuntu 10.10 on my machine as dual boot using WUBI but on a seperate partition to Windows. Loving it so far, but i cannot get any external drives to mount - i've tried pen drives, camera memory cards and hard drives but nothing comes up.
I have just tried restarting with a pen drive plugged in, and it finally showed something in the computer folder - "memory stick drive" is shown (and my internal CD drive, which i'm not sure was there before.), but i still can't access it and when I try to unmount it gives me the message
Error detaching: helper exited with exit code 1: Detaching device /dev/sdc USB device: /sys/devices/pci0000:00/0000:00:1d.7/usb1/1-1) SYNCHRONIZE CACHE: FAILED: No such file or directory (Continuing despite SYNCHRONIZE CACHE failure.) STOP UNIT: FAILED: No such file or directory
I am trying to figure out how to get the UUID for some of my external hard drives.the internet revealed a couple of promising leads, this is what I have tried so far:
ls -l /dev/disk/by-uuid -> didn't list the hard discs blkid -> didn't list the hard discs lsusb -v -> listed the hard disc but no uuid
A normally formatted usb key is listed with uuid. The external hard discs are fully encrypted by truecrypt(realcrypt). I have been reading not so great things about that itself, but for now I don't have a promising alternative that I can use with windows as well.Any google searches don't seem to cast any new light on this for me,I'd be open to suggestions if there's a better way to get a definite ID for a hard drive... I just need to be able to mount it with realcrypt