I did a search and couldn't find anything pertaining to this - if I've missed something please direct me in the right direction We have an Ubuntu box set up as a headless office server (latest desktop install of Ubuntu) and we recently set up two 1TB HDDs in a RAID1 array using mdadm - as far as I can tell it worked successfully and created /dev/md0 with an ext3 file system. After sharing the drive I can see it from the other office computers and could transfer data to and from the RAID array just fine.
I didn't figure out how to get it to automatically mount on boot so I restarted it to see if it would do so by default - however, when I restarted I couldn't see the RAID array any longer on the desktop and it came up as a 0.0kb RAID array in Disk Utility, saying it was broken. It wouldn't let me check it until I stopped and restarted the array.
After restarting I hit "check array" and it appears to be repairing the drives. What have I missed? What happened here? How can I fix it? What other info can I provide to assist:
I've got a RAID5 array that doesn't want to automount after rebooting. I'm pretty familiar with linux, RAID, and mdadm, and up until now, I've had the RAID5 array working just fine. However, whenever I reboot, the array drops off and won't remount until I manually assemble and then mount the thing. I find this odd because I had everything automounting just fine back in 10.3, and even in 11.0 (I think - not sure on that). Currently, things are working, but I'd really like to not not have to type
Code: mdadm --assemble /dev/md0 /dev/sda1 /dev/sdb1 /dev/sdc1 /dev/sdd1 followed by Code: mount /dev/md0 /mnt/data every time I reboot. Even including this in some sort of start-up script seems kludgey... Surely there must be a more elegant way of automatically bringing up a RAID5 array after booting? I'm not sure what information you'll need, so I'm going to go ahead and include as much as I can anticipate...
I created a raid1 disk with Disk Utilities on with Karmic, then upgraded my system to Lucid then Maverick. This Raid disk is just a data store, I'm not booting off of it. when I reboot the raid1 disk does not start. I have to go into Disk Utilities, stop the array and then start it. Then it comes up and I can mount it. I ran dpkg-reconfigure mdadm and it created a valid entry in mdadm.conf. but the array still does not start on boot. I want to have it auto mount with fstab but need to make sure the array starts first.
I have just upgraded to Fedora 14 from an older version. I now have problems mounting my RAID1 array, which was operating correctly until now. This is a software RAID which was initially built under Fedora 10.The array is md0, and is made of 2 SATA drives (sdc and sdd) which have only one partition. The underlying filesystem is NTFS. The array is assembled correctly and active, as reported by /proc/mdstat and mdadm -D.When I try to mount the array, I get this:
Code: [root@Goofy ~]# mount /dev/md0 /mnt/raid mount: you must specify the filesystem type
I'm sorry if this is the wrong section and if there is another thread on the matter. I searched but couldn't find threads with my specific problem. I've just installed Ubuntu 10.10 Server 64 bit which I intend to use as a internal file server.
The hdd setup is: 500gb system disk 1tb storage 2tb storage (2*2tb using built-in motherboard hardware RAID1) When the installation was complete and the computer rebooted I got an error message saying "error: no such disk". After re-installation I got the same message and I then tried disconnecting all the storage devices and it booted perfectly. I then tried connecting up the 1tb drive and again it booted as it should. But when I re-connected the RAID:ed disks the error message re-appeared.
One of my HDD on a 2 drive RAID1 md (linux software raid) is intermittently clicking. When this occurs, I can hear a loud clicking noise at uniform intervals and then it stops. Its like "click..click...click..click...click..click...click..click". You know what I mean
It does it about 4 times per hour. I believe this drive is about to die.
Until I find a replacement drive, can I run into problems with the data on the array? I believe the mdadm utility would tell me if the drive was faulty and once I replace the drive, it would auto rebuild the array (re=copy the data to the new drive)?
I have over 1.2TB of data on this array I really dont want to lose everything...
I have two HDs (let's say sda and sdb). Both are the same size and have the same partitions already (sda1/sda2/sda3 and sdb1/sdb2/sdb3). Basically they are ready to make a RAID1 array.
Writing with new udev rules, I could create and give fix HD labels with /sbin/scsi_id.
Example: For sdb1 I have a fix device name created under /dev as hd2_boot1, for sdb2 I have /dev/hd2_boot2 and finally for sdb3 I have created the device /dev/hd2_boot3.
With using the command "mdadm --create /dev/md0 --level=1 ....", I could create a RAID array.
But, when I check the status one of the RAID devices, like with the command "mdadm --detail /dev/md2", it still shows me as part of the RAID array the sdb* devices, not the hd2_boot* devices. Something like this:
I would like to see basically as member or the RAID array always the /dev/hd2_boot3 not the /dev/sdb3 (like above), is this possible?
Bottom line, I would like to keep the order of the RAID arrays depending their scsi ids, not depending their scsi numberings which is given by the kernel, since the scsi numberings (sda, sdb, sdc and etc.) can change depending the physical connection.
I'm looking to recover a RAID1 array hopefully using mdadm. Ive not really used Linux much befor but I'm keen to learn to get my data back. Basically one of the disks in my Maxtor Shared Storage II (2x500GB sata) died and I could do with either rebuilding the array or getting the data off another way.
I have a spare machine I could use for recovery process. It has a spare drive but its only 120Gig, I also have a bigger 320gig disk but thats IDE not SATA. Do I need to purchase another 500GB sata drive or can I use either of my spares? If i do need to buy a new drive could I use a 1TB or 1.5TB or will it have to be 500? Next question is what is that best version of linux to use, I have knoppix 6.2 and Ubuntu (not sure on version) already. I noticed that mdadm isn't installed by default on Ubuntu.
Like it says in the title, I am thinking it should be this hard to install the RAID1 array in my brand new PC. Here is what is happening. I have two brand new 1TB drives that I am attempting a new, fresh install of 10.10 on (in fact, the entire box is new). I am attempting to use the alternate desktop install so that I can have access to the manual partitioning (which is required to setup RAID 1, correct?).
I tried to use the guide here: [URL]... I followed the steps, but when I got the the very end (after selecting and creating the MDs) I get an error message stating that there is no root file system defined. I went back and checked all the steps and I am sure I followed everything in the guide.
Here are some quirks (not sure if they are bugs or not) In step 5 of the disc partitioning, it says to select the bootable flag and set it to yes (I am assuming). I press enter over that options, the screen flashes really quickly to a progress bar, but then comes back to the options screen and it still says bootable flag is off. No matter how many times I do it is says "off".
Also, and here is the bigger problem I think. - So the guide says to select the free space in each drive and then select Automatically Partition the free space, which I do, and it comes back and looks formatted accordingly - has 975.6 GB ext4 / and 24.6 GB swap swap. No Problem there.
BUT - whenever I do the same thing to the second drive, the partitions on the first seem to disappear. Meaning, it doesn't say free space, and has two partitions listed, but the / and the swap (last items in each row) have moved to the second drive partitions. I am not sure if this is how it is supposed to be since the pictures in the linked guide to not show what it looks like after that. THis is driving me crazy and I have to have it set up in RAID 1 and unsure as to what it is I am missing.
I installed a distro based on CentOS 5.5 (FreePBX distro FYI). It used an automated kickstart script to create an md RAID1 array of all the hard drives connected to the machine. Well, I installed from a thumb drive, which the script in interpreted as a hard drive and thus included in the array. So, I ended up with three md arrays (boot, swap, data) that included the thumb drive. Even better, it used the thumb drive for grub boot so I couldn't start up without it. I was able to mark the USB drive as 'failed' and remove from each array, and even change grub around to boot without the usb drive, but now each of the arrays is marked as degraded:
I just experienced a HDD failure and while reorganizing the drives inthis machine I realized the benefits of UUID instead of /dev/sdX nomenclature. I am trying to determine the UUID of 2 disks that are assembled in a RAID1 array. right now they are /dev/sde & /dev/sdf with each only one partition. I tried ls -l /dev/disk/by-uuid but I get only the UUID of other disks, not the ones currently ID'd as sde & sdf. my mdadm.conf assembles several raid arrays all by UUID, but somehow, I cant recall how I got the UUIDs of the other HDDs at first...
I'm not entirely a newbie, but this seems like such a simple question I'm not sure where else to ask it. I checked through the various HOWTOs and searched already and didn't find a clear answer, and I want to know for sure before we start investing in hardware. Is is possible to create a RAID1 (mirroring only) array with 3 live drives, rather than with 2 live plus a spare? Our goal is to have 3 drives in a hot-swap bay, and be able to pull and replace one drive periodically as a full backup. If I do:
I've got a mailserver set up in a raid1 array.I shut down the system to install a CD-ROM drive but forgot to change the master/slave settings (I know, don'tt say anything) and didn't realize it before Centos started booting up, so it booted the hdc drive from the array.I rebuilt the array using mdadm without any apparant issues but on subsequent bootup, I get the following error :
There doesn't seem to be any side effects to this but since that didn't happen prior, I figure there's probably something I didn't do properly since I'm fairly new to the linux world.My raid array was originally set up by the Centos instal software and is set up like this :
hda1 + hdc1 = md0 (boot) Hda3 + hdc3 = md2 (/)
The other partitions are of the same size on each drive and are swap partitions.
PS : The drive is SMART capable and no errors appear during a self-test.
edit : Clonezilla also fails to boot properly although I don't know if its due to a software raid array in the first place or the errors in the filesystem. When only one drive was detected because of the jumpers, it booted properly.
I'm convinced that mdadm is going to be the death of me. I've wasted numerous hours on this so far without luck.
OpenSuse 11.4 on an old Supermicro box, creating a software RAID1 array across 2 x IDE 500GB disks. Creating /dev/md0 as a 250MB partition across /dev/sda1 and /dev/sdd1 for /boot, another 465GB partition across /dev/sda2 and /dev/sdd2 as an LVM partition to hold volumes for the various other OS filesystems. After the initial installation and configuration there were a series of mishaps with faulty IDE cables that had drives failing to show up at boot. Somehow, /dev/sdd2 got configured to array /dev/md1 as a spare drive. And nothing I've done so far gets it to show up as an active drive.
The obvious step of failing the partition, removing it, then adding (or re-adding) will bring it back as a spare. I've tried roughly a dozen different permutations of those same steps. The latest was to 'dd if=/dev/zero of=/dev/sdd2' to clear the partition. Thought this might be the trick - after the zero, mdadm -E /dev/sdd2 reported 'no superblock' and no md1 configuration.
So 'mdadm --add /dev/md1 /dev/sdd2' and it still comes back as a spare. Here is mdadm -D /dev/md1
/dev/md1: Version : 1.0 Creation Time : Sat Jul 9 10:26:01 2011 Raid Level : raid1 Array Size : 488119160 (465.51 GiB 499.83 GB) code....
I can't stop this array, the OS is running from there. I can't easily boot from CD to repair, all IDE ports have disks attached.
Does anyone have an incantation to promote a spare to active?
I had a thread earlier asking for help installing Ubuntu Server to my homemade atom based NAS, but I gave up on that (GUI's are more helpful to me ) and so now I'm running Ubuntu 10.10 Desktop 64bit edition.Last night I was finally able to get mdadm set up. I have two 1TB disks set up in RAID1 config. I used
Code: sudo mdadm --create /dev/md0 --level=1 --raid-devices=2 /dev/sdb1 /dev/sdc1 to create the array, and then waited until it finished assembling. Then I used System>Administration>Disk
I have a used but good harddrive which I'd like to use as a replacement for a removed harddrive in existing raid1 array. mdadm --detail /dev/md00 0 0 -1 removed1 8 17 1 active sync /dev/sdb1I thought I needed to mark the removed drive as failed but I cannot get mdadm set it to "failed". I issue mdadm --manage /dev/md0 --fail /dev/sda1But mdadm response is:mdadm: hot remove failed for /dev/sda1: no such device or addressI thought I must mark the failed drive as "failed" to prevent raid1 from trying to mirror in wrong direction when I install my used-but-good disk. I want to reformat the good used drive first right? I believe I must prevent raid array from automatically try to mirror in the wrong direction.
So I didn't notice when I setup my CentOS 5.5 server that I left / as RAID 0 on md1. All the rest are RAID 1. Is there a way I can modify the array to RAID 1 without a risk of data loss? I'm glad I caught this before I setup any other services. I've only setup smb so far...
[root@ftpserver ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/md1 16G 3.0G 13G 20% /
I had some issues with my RAID6 array (with 15 disks), where 5 disks got disconnected (each five disks is connected to the motherboard via 1 SATA cable), which brought down the RAID array. I fixed this problem via readding the disks and running the array:
mdadm -R /dev/md1
However, after rebooting, the array appears inactive, and I have to go through the same motions to fix this and make it active. The array is present in the /etc/mdadm/mdadm.conf, though also 2 other raid arrays (3 arrays total):
DEVICE partitions # auto-create devices with Debian standard permissions CREATE owner=root group=disk mode=0660 auto=yes
My box has a raid5 array (mdadm) with everything in it (/boot and /) but swap that is actually spread across the 4 drives. I had ubuntu 10.10 installed (amd64) with grub1, when I upgraded to natty (11.04) it automatically installed grub2. Well boot fails, it always goes to grub rescue no matter what happens. I've installed and reinstalled grub2, and boot always fails with:
"error: file not found".
In grub rescue I can see that md0 is actually available, an "ls" to (md0)/boot succeeds but the strange thing is that an "ls" to (md0)/boot/grub prints nonsense, as does an "ls" to (md0)/boot/usr/lib/grub/i386-pc/. When I try to load the required modules for boot (linux raid etc) in grub I also always get a "file not found error" (I fsck'd md0, which says everything's fine). I have installed the latest version of grub2 and executed grub-install in all four drives.
I just tried to install with Ubuntu 10.04 AMD64 Alternate on RAID1 and Encryption but after reboot the screen just stays black.
my system is a AMD Athlon 64 X2 Dual Core 4200+ on a Abit AN-M2HD Motherboard, and 2 HDs each 250.1 GB i split the HD into * 50GB for / * 200GB for /home * 1GB for swap all get a RAID1 /home is encrypted with passphrase (Twofish 256, cbc-essiv:sha256) swap is encrypted with random (Blowfish 128, cbc-essiv:sha256)
where can i check RAID and hardware compatibility?
1st I am relatively new to linux (but not to *nix). I have 4 disks assembled in the following intel ahci bios fake raid arrays:
2x320GB RAID1 - used for operating systems md126 2x1TB RAID1 - used for data md125
I have used the raid of size 320GB to install my operating system and the second raid I didn't even select during the installation of Fedora 14. After successful partitioning and installation of Fedora, I tried to make the second array available, it was possible to make it visible in linux with mdadm --assembe --scan , after that I created one maximum size partition and 1 maximum size ext4 filesystem in it. Mounted, and used it. After restart - a few I/O errors during boot regarding md125 + inability to mount the filesystem on it and dropped into repair shell. I commented the filesystem in fstab and it booted. To my surprise, the array was marked as "auto read only":
and the partition in it was not available as device special file in /dev:
I have a 10x2tb disk array that i'm trying to build into a single software raid 5 i have tried this 2 times now the first it made it to 58.7% and the machine locked up and the array would not restart after a reboot. On my 2nd try all was looking good until about 50% i noticed that the speed dropped in 1/2 and that ksoftirqd/2 is taking up a lot of cpu (about 90%) the md0_resync and md0_raid5 are also taking 60-90% when the build started they took 7%. when i do a dmesg i see a lot of the message compute_blocknr: map not correct.
For a little info on the physical setup this is running on an Atom 510 with 2GB of mem the drives are connected to an addonics 4-Port RAID 5 / JBOD SATA II PCI Controller using the sil3124 chipset. I'm using 2 addonics 5X1 SATA Port Multiplier connected to the controller to get the 10 drives attached. All drive show up and don't seem to have any issues. I'm running a fully updated as of 3/20/10 version of centos 5.4
I will let this continue to run over night but i expect it to be locked up by morning if it follows what the last attempt did.
When I try to use Wubi to install Ubuntu it fails on reboot. It seems to run a script on reboot then nothing. I get a command prompt window with the script and the laptop seems to lock up. Only a hard re-boot will restart the system. I do get a dual screen boot prompt Windows/Umbutu but if I choose Umbutu, I get the same command prompt screen and the system is locked up again. Chosing Windows at the boot screen works fine.
I'm installing 10.4 x64 on a Dell PowerEdge R610 RAID 1 and the install goes fine until I have to reboot the finish it. On reboot after about a minute it gives the error
Code: Gave up waiting for root device. Common problems: -Boot args (cat /proc/cmdline/ -check rootdelay= (did the system wait long enough?) -check root= (did the system wait for the right device?) -Missing modules (cat /proc/modules; ls /dev) ALERT! /dev/mapper/mysql1-root does not exist. Dropping to shell!
what to do and my co-worker who knows a little more about Linux can't get it to let him check any of the above common problems.
We're tried letting GRUB manually configure itself at the end og the install before we reboot and we've also pointed it to /dev/sba partition which is what should be the boot partition.
EDIT: in the code tag it's ALERT! /dev/mapper/mysql(one)-root.... the L and 1 look the same
I'm installing 10.4 x64 on a Dell PowerEdge R610 in a RAID 1 configuration and the install goes fine until I have to reboot the server to finish the install. On reboot after about a minute it gives the error.We've tried letting GRUB manually configure itself at the end of the install before we reboot and we've also pointed it to /dev/sba partition which is what should be the boot partition.In the code tag it's ALERT! /dev/mapper/mysql(one)-root.... the L and 1 look the same.I've got 3 of these server to get setup and running as part of our back end and network restructuring and have been stuck at this point on the first one since Friday.
I recently installed Ubuntu 10.04.1 LTS and have it set to dual-boot with windows 7 on my gateway tablet PC. Since installing, sometimes when I suspend or reboot, my wireless does not work. To be more clear - it is working fine, then I either suspend/reboot, then I come back and no wireless networks show up at all. I can still connect with ethernet cable though. I can solve it by rebooting multiple times (usually around 1-5 times) and then it (magically) seems to work again. I have also tried switching from network-manager to wicd, but the problem persists. With both network-manager and wicd, I've tried simulating a reboot for the wireless by using the command 'sudo modprobe -r iwl3945 && sudo modprobe iwl3945' (network-manager) and 'sudo /etc/init.d/wicd restart' (wicd), but these do not seem to work.
Just installed Ubuntu a month ago (installed within windows), and for the 2nd time it has gotten stuck in grub upon restart after doing an automatic update. Last time, I reinstalled from the CD, but I can't do that every time. Learning how to start up from grub would probably be worth learning so I don't have to use the CD. But fixing the problem is most important. It makes me wonder if it was a mistake to install Ubuntu from within Windows (does this mean I have "wubu"?). I did find a webpage on booting Ubuntu from grub, but I could not get the kernel to load, regardless of if I used the instructions for wubu or not.
I can get connected to the Internet when I reboot the system, but the connection usually fails after within 10 mins. However, even then the network manager says it's connected. Never reconnect, manually or automatically.