Ubuntu Servers :: HW RAID Disk Shows Up In Fstab But Not In /dev/disk/by-uuid?
Jun 28, 2010
I have an SiI hardware SATA RAID card, with two 500GB disks in mirrored RAID configuration. When I first plugged them in and set it up, things seemed to work ok, but on boot the raid controller told me that the RAID needed rebuilding, and it would happen automatically after POST. So I didn't worry about it, and the drive mounted fine, and it's been that way for years. I just went in and manually on-line rebuilt the RAID in the controller's BIOS, and now when I boot into Ubuntu, both disks show up in fdisk, but neither show up in /dev/disk/by-uuid. Am I missing something?
View 9 Replies
ADVERTISEMENT
Jan 5, 2010
What would be the best way list disk and partitions in the fstab file?
View 1 Replies
View Related
Aug 1, 2010
I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.
I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).
These are the commands I used:
Quote:
p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*
[Code]....
I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...
So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.
It's still trying to use /dev/disk/by-uuid/412d... to boot.
What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.
View 2 Replies
View Related
Oct 16, 2009
my Fedora 11 system is not starting anylonger. It stops with the message:
Code:
VFS: Can't find ext4 filesystem on dev dm-0
The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).
View 2 Replies
View Related
Mar 12, 2011
I'm testing my ability to recover a failed disk on a three disk software RAID 5 setup.
I have used a 10.04 alternate install disk to setup a three disk RAID 5 array according to this: [URL]. This is for a RAID1 setup. I followed it exactly except that I performed the steps on three drives rather than two and selected RAID5 instead of RAID1. Each disk is 500GB and has a 26 GB swap partition and the remaining space on each disk set as / with the boot flag on.
I installed the OS on my array and everything boots without a problem. After I booted up I started a terminal and ransudo dpkg-reconfigure mdadm to set the boot degraded to true and rebooted.
Next, I shut down the computer, disconnected the power on drive 1 (sdb) and then tried to boot. I get this (not verbatim):
Quote:
mdadm: CREATE user root not found
mdadm: CREATE group disk not found
raid5: raid level 5 set md0 active with 2 out of 3 devices, algorithm 2
mdadm: /dev/md0 has been started with 2 drives (out of 3)
[Code].....
*then a list of common problems and then:
ALERT! /dev/disk/by-uuid/bunchanumbersnad letters does not exist. Dropping to a shell
Then it dumps me to initramfs. MD0 is the swap partition. At this point I don't know what the heck to do. I'm skating on the edge of noobidity and this is pretty much over my head.
I want to use this server as a virtual machine server and the desired behavior would be that, if a hard drive should fail, the server would alert me via email and continue to run in a degraded state.
Is it even possible to install the OS on the array and run it degraded? Given the desired behavior, should I be looking at something other than RAID5? My client is broke so I'm trying to avoid a hardware RAID if I can do it.
View 1 Replies
View Related
Apr 14, 2010
One of my raid disks (was a software raid5 with 4 drives) failured. I wanted to buy a new 1.5TB harddisk anyway so i i copied all the Raid data onto the new disk and disconnected the old ones. But then the new Disc crashed right before i could mirror it with another 1.5TB disk. So i need to reassemble my old Raid 5 now. I connected the drives in the former order except the first one, because this was the failed disk. The problem is, mdadm can't find the raid, no superblocks. Fdisk doesnt show mdraid superblocks, too. But fdisk has never showed superblocks. The thing is, this raid was crypted, but i crypted the /dev/md0, so this doesn't affect anything here, or?
some infos:
Code:
server:~# mdadm --misc -d /dev/sda
mdadm: option -d not valid in misc mode
server:~# mdadm --misc -t /dev/sda
[code]....
Disk /dev/sdc doesn't contain a valid partition table But fdisk always said there is no valid partition table and the raid was working.
View 4 Replies
View Related
Nov 23, 2010
My raid array has failed. I have two disks /dev/sda and /dev/sdb./dev/sdb has failed and I could not rebuild the array(madm returned that the device is busy) so I rebooted the machine. After that, the whole sdb disk went missing, as it now only shows sda in fdisk -l.Did the disk went totally dead or my raid glitched?
View 8 Replies
View Related
May 5, 2011
My system locked up while copying files last night. My RAID array will not start. I did verify my UUID's. (Lesson learned.) I do not understand a few things.1. Why do different drives show "active sync" on different drives? 2. Why does "Disk Utility" tell me the RAID is not running and when I try to assemble the RAID, mdadm returns: mdadm: device /dev/md0 already active - cannot assemble itWhen I try to start the RAID using "Disk Utility":
Code:
Error assembling array: mdadm exited with exit code 1: mdadm: cannot open device /dev/sdd1: Device or resource busy
mdadm: /dev/sdd1 has no superblock - assembly aborted
So, I examine sdd1:
Code:
sudo mdadm -E /dev/sdd1
[Code]...
View 9 Replies
View Related
Jun 7, 2010
I'm looking to set up a bit of a home server, and am wondering about storage. What I'd like is something like RAID 6, which has good redundancy built-in, but with this being a home server, I'd prefer to start a little smaller and leave room to build it up in future. I'd been looking at commercial products like the 'drobo', which seems fairly ideal, but I'd really like to see if I can do it myself. I understand that throwing the RAID into an LVM will allow for some expansion, but the last time I checked, most RAID setups called for the same sized disks, or at least limited the array by the size of the smallest disk present.
What I'd like is the ability to build a basic framework with a few cheap disks, and then as things start filling up, to be able to add larger ones (perhaps eventually pulling out smaller ones as though they'd failed and replacing them with big ones)
View 1 Replies
View Related
Jan 9, 2011
I've got a raid5 array of 4 disks with ubuntu 8.04 runing on it that is currently still working:
/dev/sda
/dev/sdb
/dev/sdc
/dev/sdd
Smartmontools for /dev/sdc tell that there are 9 sectors pending for reallocation:
Code:
197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 9
198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 9
And /dev/sdd has increasing number of reallocated sectors (about 1 every couple of minutes):
Code:
5 Reallocated_Sector_Ct 0x0033 100 100 036 Pre-fail Always - 1735
/dev/sdc has failed a coulple of times this week (but I have always sucessfully readded it to raid5) . But the increasing number of reallocated sectores on /dev/sdd concerns me even more.
I'm affraid that during removal of /dev/sdd and adding new /devs/sdd disk, raid might fall appart. That's why I would try to do it in Ubuntu Live CD:If the raid falls appart (/dev/sdc fails) during the readding of new /dev/sdd disk, I might still remove the new /dev/sdd and return the previous one and assemble the raid with:
/dev/sda
/dev/sdb
/dev/sdd (old one that was previously removed)
Does assembling Raid in Ubuntu Live and adding new disk for /dev/sdd write anything on /dev/sda, /dev/sdb and /dev/sdc in the process of adding /dev/sdd into raid5?
View 2 Replies
View Related
Jan 15, 2010
I've had two hd's in my box forever. for more space and backup reasons. Well I have started running the Debian Squeeze distro since December. I've had many issues, some are still unresolved. but now I'm running into major headaches with the fstab. Specifically dealing with/wondering why UUID's are used instead of the old /dev/hd? I was a little annoyed when I tried Kubuntu to find /dev/sd? used instead of /dev/hd? but that was workable. But the UUID's are a nightmare. Here's my problem.
My main box is finally giving up the ghost. The mobo is dying. So in order to do some tests I took my hd bundle (my two hard drives with their cables) physically out of the box and temp installed them in a test box. I wanted to do some benchmark and other tests. I got all kinds of errors. I found that the system wasn't recognizing the UUID's listed in fstab. My concern is when the new mobo gets here next week I won't simply be able to plug the hd's in like I always have been and just let Linux reconfigure itself (Debian used to be good about this). I really don't want to have to clean reinstall if it's not needed.
So for this I have two questions. WHY developers decided to drop using /dev/hd? or even /dev/sd? ?
And is it possible to revert fstab's listings back to the old /dev/hd? settings. In debian fstab had lines commented out showing how each partition was listed in it's /dev/hd? status during install.
I'm getting really sick of all these archane changes in ALL aspects of linux that don't seem to have any good explaination or need.
View 4 Replies
View Related
Oct 4, 2010
Is there any C function that will translate UUIDs into device names? I have a little graphical mount tool that can read user-mountable device names from /etc/fstab and lets you cycle through the list and mount or unmount them. But it doesn't work with UUIDs, which are preferred these days. Is there any way around this?
View 2 Replies
View Related
Jan 23, 2010
Fresh install of 9.10. Update all. The update froze at the very end: configuring grub-pc. I let it run like that all night before settling for a reboot while the update manager was still running. (The update also raised the kernel version.) After reboot, I tried the newer kernel [2.6.31-17-generic] at the grub menu and got this error:
Quote:
udevadm trigger is not permitted while udev is unconfigured.
udevadm settle is not permitted while udev is unconfigured.
svgalib: Cannot open /dev/mem.[code].....
And, cat /proc/modules gives me a long list of modules (too much to retype for now).So my hard drive is there. maybe some errors on the / home partition [sda6] and I can get to the root directory.
FYI: MS-win XP still boots fine from grub.
View 1 Replies
View Related
Feb 6, 2010
I run a headless Ubuntu 8.04 server, which acts as a web, email and file server. I am sticking with 8.04 as it is a LTS release and will upgrade to the next LTS when it is released.
I have two external USB drives, that I need to mount at boot. I have been using /etc/fstab up until now, with the following entries:
Code:
However, as I gather from doing searches is quite common, occasionally I get an error during boot (causing the system to drop to a recovery shell) because the USB drives take time to wake up and the system hasn't found them by the time it reads /etc/fstab.
From doing searches, it seems there is nothing you can do to fstab to fix this, so you need to mount them using an rc.local script instead, using:
Code:
The problem is, as I have two USB drives, their /dev/sdxx location changes between boots. I thus want to use UUID codes as I do in fstab, however I haven't found anything about this.
Does anyone know how I can use the mount command and UUID to mount a drive in rc.local and what options I have to use the mount the drive with the same options that I am using in my fstab entry? Obvisouly, I can't refer back to fstab using the mount command, because then I will still get the boot error issue if they are listed in fstab. And there is no space internally for the USB drives as there is already two internal drives.
View 3 Replies
View Related
Apr 13, 2010
I want to use Access Control Lists (ACLs) on a removable usb hard drive. I don't know how to set up /etc/fstab for usb drives. Every time I try to make an fstab entry for the usb drive I get an error, when I plug in the usb drive and the system tries to automatically mount it. The drive isn't mounted. If I delete the fstab entry for the usb drive, the drive automatically mounts with no problem. However, ACLs are not enabled, because no fstab entry exists to enable ACLs.
The error message states that "only root can mount the drive". However, as far as I can tell, automatic usb mounting is being done by root. When I plug in the drive (with no fstab entry) and it automatically mounts, Nautilus Properties shows the drive is owned by root and has permissions of rwxr-xrwx.
View 1 Replies
View Related
Sep 27, 2010
I have a NETGEAR ReadyNAS NV+ with four 1TB drives in a RAID-5 array. This is our primary file storage. This has previously been backed up to a hardware RAID-0 array directly attached to our Windows server. The capacity of this backup array is no longer sufficient. So the plan was, take a bunch of 200GB to 320GB drives (And a 750) I had kicking around, chuck them in a couple of old SCSI drive enclosures I have collecting dust, attach them via IDA/SATA-to-USB adaptors to a USB hub, attach that to the server, create a JBOD array spanning the disks, and back up the NAS to that. Performance is not an issue as this is just to be used for backup, with the idea being as near to zero cost as possible (Spend so far = NZ$100�ish).
The first hurdle I struck was Windows not supporting Dynamic Disks on USB drives (Required to create a spanned volume). At first I resisted using another machine (i.e. a machine running Ubuntu) as I didn't want to dedicate a piece of hardware to backing up the NAS. I then decided it would be acceptable to do this via a VM, which is what I've done.So I have 10.04 running under VMWare Server 2.0.2 under Windows Server 2008 R2. The disks are all presented to the VM. I wasn't sure if I was going to end up creating the array under LVM or something else, but I noticed Disk Utility has an option to create an array, so I tried that. When I add two 250GB drives, the array size is 500GB. When I then add a 160GB drive, the array size drops to 480GB. Huh? If I keep adding disks (Regardless of order) the final array size comes out at 1.8 TB, as per the attached screenshot. Now with the following drives, I expected something more like:
160 + 250 + 250+ 750 + 250 +200 + 200 + 250 + 320 + 250 + 320 = 3.2TB
Am I missing something or making a false assumption somewhere?
View 4 Replies
View Related
Mar 18, 2010
I'm running Kubuntu Karmic on my Dell Inspiron laptop - about 200 bug fixes behind because my only available internet is a cellular connection on a crappy wi-fi router - and last night, I suspended it, but it shut down instead. Not a problem, it does this fairly often, figure the RAM gets jostled or something.
But when I go to boot it up, it gets stuck at the pre-loading screen before getting garbled and dropping to the shell, where it says "mount: mounting /dev/disk/by-uuid/[insert hex code here] failed: invalid argument". Of course, mounting /root/sys, /root/dev and /root/proc fails, (directory does not exist) and it gives me the busybox initramfs prompt.
View 6 Replies
View Related
Oct 16, 2010
Basically, today I installed Ubuntu on my Toshiba NB200 to dual boot with XP. The installation went okay (I think) but after, when I rebooted, and selected Ubuntu from the possible selections, I got this message:Quote:Gave up waiting for root device.Common problems:
-Boot args (cat /proc/cmdline)
-check rootdelay= (did the system wait long enough?)
-check root= (did the system wait for the right device?)
[code]...
View 3 Replies
View Related
May 26, 2010
I was able to successfully create a small, fixed-size "ram disk" just for kicks, via:
Code:
sudo mkfs -t ext3 -q /dev/ram1 65536
sudo mkdir -p /media/ramdisk
sudo mount /dev/ram1 /media/ramdisk -o defaults,rw
However, I wanted it to auto-boot, so I added the line to fstab:
Code:
/dev/ram1 /media/ramdisk ext3 defaults 0 0
This, however, rendered Ubuntu unbootable and I had to repair the fstab by removing the line via a Live CD. I had enough patience to boot through the CD an extra time to see if just deleting "ext3" would work, but it did not. What would cause this to make the system unbootable and how I could make it work?
View 5 Replies
View Related
May 18, 2011
ASUS eeepc900 Easy Peasy 1.6 (Ubuntu 10.04)
After I boot the 16GB second SSD disk is mounted as HOME on /media This seems to be a default. However there is no entry in /etc/fstab to do this. How does it happen?
I want to move my home directories to be on the 16GB disk so I need it to be mounted on /home rather than where it is by default which is /media/HOME
View 3 Replies
View Related
Mar 15, 2011
One entry I have put in fstab results in the failure of a partition to be mounted at boot time. I get the message:
Code: The disk drive for /media/WinXP is not ready yet or not present. Continue to wait; or Press S to skip mounting or M for manual recovery If I choose M and enter the command: Code:mount -t ntfs /dev/sdb1 /media/WinXP then I get no error message, but the partition still doesn't seem to be mounted, when boot completes.
I don't understand this failure. I have created my fstab file using UUIDs to boot Ubuntu on my dual boot machine. It works fine, booting from the hard-disk which is Master on my Secondary IDE channel. For Ubuntu booting the MBR and grub menu are on this disk. The default is to boot Ubuntu , but with an option to select Windows Xp.
As an aside, I can set an option in my BIOS to make the Master disk on the Primary IDE channel the first disk, rather than the second disk. Then the system boots from the MBR on this Primary IDE channel and boots only to WinXP. That works fine.
When running Ubuntu I use space on the Windows disk (on the Primary IDE channel) to hold backups of key Ubuntu files in case I loose Ubuntu - as I did for the past few days. So, to mount this partition I inserted this line into my fstab:
Code:
UUID=0e4851c44851ab6b/media/WinXPntfsnosuid, nodev, allow_other00 I know the UUID is correct because I have checked it with blkid. But the partition is not mounted at boot time. I don't even get an icon for the partition on my desk top. It appears in the 'places' menu, as unmounted, but mounts as soon as I click on it. However, this causes some of my linux apps, which want to load and save to this partition, to post an error message until I have manually mounted it via clicking on it in the Places menu. I want to avoid this manual step by having the partition automatically loaded at boot time. What am I doing wrong?
View 9 Replies
View Related
Jul 4, 2010
I've got a removeable disk which I want to mount on startup automatically at mountpoint "/backupsystem". If' it's not there I would like to have no error message. Actually after upgrading to 10.4 I get the message: Continue to wait; or press S to skip mounting or M for manual recovery.". But I don't wand this if the disk is not there that's OK for me. How would I configure fstab to achieve this?
View 2 Replies
View Related
Feb 5, 2010
I am running SUSE 11.1 on a 80Gig IDE HDD, I have added a 160 SATA HDD which I wish to use as storage.
fdisk reports it as /dev/sda1 - W95 fat 32 LBA. What would the fstab entry be to make the disk mount automatically on boot, so that it shows on the desktop ready for use.
View 5 Replies
View Related
Jun 30, 2011
I know you can fail and then remove a drive from a RAID5 array. This leaves the array in a degraded state.
How can you remove a drive and convert the array to just a regular, clean array?
View 9 Replies
View Related
May 28, 2010
I'm wondering why the NetworkManager log file would show an entry whenever an external usb drive is plugged in. It's obviously HAL related but looking at the various files of hal, udev, syslog shows nothing relevant.
Here's the entry just after the usual info about the network card:
May 28 08:35:16 doomed NetworkManager: <info> Activation (eth0) Stage 5 of 5 (IP Configure Commit) scheduled...
May 28 08:35:16 doomed NetworkManager: <info> Activation (eth0) Stage 4 of 5 (IP Configure Get) complete.
May 28 08:35:16 doomed NetworkManager: <info> Activation (eth0) Stage 5 of 5 (IP Configure Commit) started...
[Code].....
This is an opensuse 10.3 with final updates system.
Any search here about the log shows nothing but actual network problems.
View 5 Replies
View Related
Dec 30, 2010
After installing the RT kernel, and updating my boot loader, I get this message. Code: ALERT! /dev/disk/by-uuid/... does not exist. Dropping to a shell! It doesn't make a difference whether I pass the 'root=/dev/', or 'root=UUID=', options to the kernel. I've also noticed a message while the system was attempting to boot up. Code: host side 80-wire cable detection failed, limiting max speed to UDMA33 This is all strange to me as I was running the Debian 2.6.32-5-686 kernel, without any problems.
View 1 Replies
View Related
Jul 17, 2010
point me in the direction to get a step by step guide to setting up a Raid 5 using the Disk Utility and 3 spare drives? I have the main OS files on a 80gig drive and I would like to mount the 3 drives as Raid 5.Just shooting in the dark now.. Screen shot is attached. [URL]...
View 1 Replies
View Related
Jul 20, 2011
Has anybody ever used Disk Utility to set up software RAID? Here I am running terminal commands (I'm a terminal junkie) and I just happen to stumble across instructions that indicate "Or you can just set it up through Disk Utility."
Sure enough in disk utility, it looks like all of the configurable options are there. It makes me wonder, though... is this kind of GUI functionality something that isn't really solid? Or does it operate predictably and effectively?
View 4 Replies
View Related
Oct 23, 2010
I am running an Opensuse 11.2 system as a backup file server. Several external hard drives are attached via USB and mounted during boot via the fstab file. Recently, I noticed the df command showed the root partition being >65% full, while the du command showed the same drive was only about 20% full. After much investigation, I discovered that some large directories had been created under the /media folder on the root partition that were hidden when the external drives were mounted as directories with the same names under /media. I could only see this when I rebooted with a live CD and mounted the suspect partition without the external drives mounted. Is there a way to examine whether a mounted partition has existing files "hidden" when a separate file system is mounted "on top of it"?
View 9 Replies
View Related
Jul 8, 2011
Before doing a clonezilla project I opened Disk utility shows the first partition labeled as sdb2, then second partition as sdb1is this normal? I will add that this is a windows drive, but I wanted to back it up before installing debian to it. How will the disk partition labeling affect partition naming in debian?
View 2 Replies
View Related