after a failed upgrade from 9.10 to 10.04 I had to format my computer and do a clean install of 10.04, and now my mdadm raid5 array wont start.my array is called "The Library", and i believe the space between "The" and "Library" is causing the command disk utility uses to start the array to fail.The exact error isAn error occurred while performing an operation on "The Library" (RAID-5 Array): The operation failed
Error assembling array: mdadm exited with exit code 1: mdadm: unrecognised word on ARRAY line: Library
mdadm: unrecognised word on ARRAY line: Library
I have a problem with my mdadm RAID. I wanted to know if anyone had any experience with shrinking RAID5 arrays. I was growing the array from 5 to 6 devices however the grow got interrupted and it has recovered to 5 drives. The 6th drive is toast and I am unable to re add it to the system. I would like to drive the device listed as "removed". I have tried mdadm /dev/md0 --remove detached and failed with no success. I am running Ubuntu kernel 2.6.28-11 and mdadm is v3.1.1.
Here is output of a "mdadm -D dev/md0" /dev/md0: Version : 0.90 Creation Time : Wed Jan 12 00:46:41 2009 Raid Level : raid5 Array Size : 4883812480 (4657.57 GiB 5001.02 GB) Used Dev Size : 976762496 (931.51 GiB 1000.20 GB) Raid Devices : 6 Total Devices : 5 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Mon Feb 15 20:25:07 2010 State : active, degraded Active Devices : 5 Working Devices : 5 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 64K
UUID : 74fa5199:84b88e81:4ae0fbae:92643084 Events : 0.1331010 Number Major Minor RaidDevice State 0 8 16 0 active sync /dev/sdb 1 8 32 1 active sync /dev/sdc 2 8 48 2 active sync /dev/sdd 3 8 0 3 active sync /dev/sda 4 8 64 4 active sync /dev/sde 5 0 0 5 removed
I have created a RAID 5 array using the built in Disk Utility.This is great and formatted it with ext4, and mounted it.However on reboot in Disk Utility as RAID Array is not running and under state Not running, partially assembled.I have to stop the array then restart it, then mount it before I can access what is on it.Obviously this is not very good as I often have the system shutdown at night to converse energy, and having to do this every time it boots is a pain.Could someone please explain in plain english what I need to go to get my array to start and mount on startup
I am having a RAID5 with a spare(total 4 disk).then the steps which lead me to a problem:
1. i was doing I/O on the array. 2. i pulled out a drive manually. So the spare drive took care of the failed one and started rebuilding. then 3. in the, mean time i pulled out the power plug of my NAS box. 4. After power up i saw my array was not active(by -D command option of mdadm). then 5. i executed: mdadm --assemble --scan /dev/md0 it gave me
I checked into the linux source and found that bd_claim is a function inside fs/block_dev.c and it failing due to which lock_rdev function (calling bd_claim in md.c) is failing and we are not able to start the array.I don't know why my RAID is not live after power on. Plese help atleast can i save my data?
This is message I get when I try and start itmdadm: /dev/md0 assembled from 2 drives - not enough to start the arrayBelow is the information I've collected about any help on how I can get the raid back up and going to I can get the data off of it would be awesome
I have a RAID 5 array, md0, with three full-disk (non-partitioned) members, sdb, sdc, and sdd. My computer will hang during the AHCI BIOS if AHCI is enabled instead of IDE, if these drives are plugged in. I believe it may be because I'm using the whole disk, and the AHCI BIOS expects an MBR to be on the drive (I don't know why it would care).
Is there a way to convert the array to use members sdb1, sdc1 and sdd1, partitioned MBR with 0xFD RAID partitions?
I powered down my file server this morning and started it back up a little while ago. One of the drive in a raid 6 array hung the system if it's inserted so after a little tomfoolery to identify the trouble drive. After removing it I finally have it booting but the raid array come up as inactive and all spares.I've "mdadm --examined" all the remaining drive in the drive in the raid6 array and the "events" are all the same, "2085124".When I assemble the array I get
120GB Sata HDD - Primary OS drive 3 x 1.0TB Sata HDD - Raid 5 array
This is on a C2D MSI P35 Platinum board. Anyway, did a fresh install of F12 on the 120GB, which I had problems with - Anaconda refused to see the drive. Fedora Live could see it fine, and it was listed as an 'nvidia_raid_member' - no idea why, but I completely erased the disc under the Live CD and proceeded to install F12.
Once F12 was installed, I loaded up mdadm to re-activate my Raid 5 array, using 'sudo mdadm --assemble --uuidthe uuid) - and it started with only 2 of the 3 drives. My /dev/sdb drive did not activate into the array, due to what mdadm said was a mismatched UUID. Ok, so I erased /dev/sdb, intending to rebuild the array. Erased /dev/sdb, and then attempted 'sudo mdadm --add /dev/md0 /dev/sdb' and I get this error: "mdadm: Cannot add disks to a 'member' array, perform this operation on the parent container" - I can find NO information on this error message.
I don't believe the hard drives are connected in the exact same order they were in before - I disconnected everything in the system and blew it out (it was pretty dusty)
Created my own file server/nas, but get stuck in a problem after couple of months. I have a server with 4x 1,5tb disks, all connected to sata ports and 1 40gb ata133 disk running ubuntu 9.10 x64 amd. I've created a raid5 array using mdadm. It all worked great for couple of months but lately the raid5 array is degraded. disk sdd1 is faulting every few days. I have checked the drive but it is fine. If I re-add the disk and wait for 6 hours my raid5 array is all fine again, but after a few shutdowns, it is degraded.
my mdadm detail:
root@ubuntu: sudo mdadm --detail /dev/md0 /dev/md0: Version : 00.90 Creation Time : Mon Dec 14 13:00:43 2009 Raid Level : raid5
I have ubuntu server 10.04 on a server with 2.8ghz 1gb ddr2 with the os on a 2gb cf card attached to the IDE channel and a software raid5 with 4 x 750gb drives. On a samba share using these drives I am only getting around 5 MB/s connected via wireless N at 216mbps and my router and server both having gigabit ports. Is a raid 5 supposed to be that slow? I was seeing speeds of anywhere from 20-50MB/s from other people and am just wondering what i am doing wrong to be so far below that.
I cant seem to get my RAID 5 (consisting of 8 1tb hard drives) assembled for some reason and I have no idea why and cant find any solutions online. Ill go ahead and show what my problem is:
here is all my hard drives:
Code: server:~$ sudo fdisk -l Disk /dev/sda: 10.2 GB, 10242892800 bytes 255 heads, 63 sectors/track, 1245 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk identifier: 0x0004f041
So as you can see the array for those last four look fine however for the first four it marks the last four drives as faulty for some reason. I am kind of clueless to do from this point on honestly, I have data on this array that I'd really like to save.
My fileserver initially had 3 1TB drives in RAID 5 configured with mdadm as /dev/md1. (System root is a mirrored raid on /dev/md0) I went to go add a 4th 1TB drive to /dev/md1 and grow the raid 5 accordingly. I was initially following this guide: [URL] but ran into issues on the 3rd and 4th commands. I've been trying a few things to remedy the issue since, but no luck. The drive seems to have been added to /dev/md1 properly, but I can't get the filesystem to resize to 3TB. I also am not entirely sure how /dev/md1p1 got created, but it appears to be the primary partition on the logical device /dev/md1. Relevent information:
The filesystem originated as ext3, I believe its showing up as ext2 in some of these results because I disabled the journal when doing some initial troubleshooting. Not sure what the issue is, but I didn't want to blindly perform operations on the filesystem and risk losing my data.
I am planning on setting up a 4x1TB RAID5 with mdadm under Ubuntu 9.10. I tried installing mdadm using "sudo apt-get install mdadm", all worked fine except for the following error: Code: Generating array device nodes... /var/lib/dpkg/info/mdadm.postinst: 170: /dev/MAKEDEV: not found failed. The end result is the /dev/md0 device has not been created, as can be seen here:
Code: windsok@beer:~$ mdadm --detail /dev/md0 mdadm: cannot open /dev/md0: No such file or directory After googling, I found the following bug which describes the issue: [URL] However it was reported way back in April 2009, and it does not look like it will be fixed any time soon, so I was wondering if anyone knows a workaround for this bug, to get me up and running?
I am getting really frustrated with trying to get my RAID5 working again. I had a RAID5 array built with 4 of the Western Digital 1.5tb "Advanced Format" drives, WD15EARS. However, when copying 1.5gb dvd encoded files to the drive, I was getting speeds of ~2mb/s. When researching how to make this faster, I came across all the posts about the Advanced Format drives and how that was causing a lot of issues for a lot of people. It looked like the solution was simple enough: partition starting at sector 64 or 2048 or whatever and then recreate the RAID. However, this is not working for me.
Here are my computer specs: Motherboard: Gigabyte GA-EP43-DS3L LGA 775 Intel P43 ATX CPU: Intel Core 2 Duo E8400 Wolfdale 3.0GHz 6MB L2 Cache LGA 775 65W RAM: 4gb DDR2 1066 (PC2 8500) Video card: ASUS GeForce 9600GT 512MB 256-bit Linux: 10.04
i was adding another disk to my raid 5, all was going well it started the reshape, got past the critical zone, worked for 20mins, but now it seems to have crashed.When i cat /proc/mdstat, or mdadm -D /dev/md0, those programes hang and dont print anything or return.from my kern.log i can see that there was an error on a disk, the raid array removed it, was going to continue the reshape but finished immediately. Anyone know what i should do?
Mar 6 16:20:26 Aries kernel: [1931119.599107] md: reshape of RAID array md0 Mar 6 16:20:26 Aries kernel: [1931119.599107] md: minimum _guaranteed_ speed: 1000 KB/sec/disk. Mar 6 16:20:26 Aries kernel: [1931119.599107] md: using maximum available idle IO bandwidth (but not more
I've been playing with this for hours, and have been unable to figure it out. I tried to convert my RAID5 array of 4 active disks and 1 spare to a RAID6 with 5 active disks.
I did this:
Code: mdadm --grow /dev/md4 --raid-devices 5 --level 6 Here is what I have on /dev/md4:
Code: /dev/sde1 active /dev/sdg1 active /dev/sdj1 active /dev/sdf1 active removed /dev/sdh5 spare code....
but it tells me that /dev/sde is busy, and then that it has a bad superblock (From what I've read, I'm sure the bad superblock is just because of the "busy" message). I've tried this with the -f option, too, with no luck.
Something weird happened last night and my raid5 failed. I am trying to re activate it and see if my data is dead or what. When I run mdadm -Asv /dev/md0 I get
Code: mdadm: looking for devices for /dev/md0 mdadm: cannot open device /dev/dm-1: Device or resource busy mdadm: /dev/dm-1 has wrong uuid. mdadm: cannot open device /dev/dm-0: Device or resource busy mdadm: /dev/dm-0 has wrong uuid. mdadm: cannot open device /dev/sde2: Device or resource busy mdadm: /dev/sde2 has wrong uuid. mdadm: cannot open device /dev/sde1: Device or resource busy mdadm: /dev/sde1 has wrong uuid. mdadm: cannot open device /dev/sde: Device or resource busy mdadm: /dev/sde has wrong uuid. mdadm: cannot open device /dev/sdd: Device or resource busy mdadm: /dev/sdd has wrong uuid. mdadm: cannot open device /dev/sdc: Device or resource busy mdadm: /dev/sdc has wrong uuid. mdadm: cannot open device /dev/sdb: Device or resource busy mdadm: /dev/sdb has wrong uuid. mdadm: cannot open device /dev/sda: Device or resource busy mdadm: /dev/sda has wrong uuid.
Relatively inexperienced user using Linux/ubuntu. Not too savvy I admit and like to use GUI as much as possible. Not a great fan of the Terminal window... I have installed a couple weeks ago Ubuntu 10.10 (Desktop Edition) using Alternative install disk (don't ask why!) on 4Gb usb stick. Working fine except one thing with the raid array. I have created a raid5 array made of 6 drives using GUI (Disk Utility). After an expansion of the array (or was it a reinstall of the OS, I can't remember exactly?) the array does not autostart anymore. Of course nor does it automount anymore.
THE WEIRD THING is that I can still start it MANUALLY from the "Disk Utility" GUI after two tries. And it works just fine thereafter!!! The first time i try to start it gives an error (something about /dev/md0_127 being not ready or buisy). THE SECOND TRY ALWAYS WORKS like a charm, the array starts and i can mount it just fine. Here is a screenshot: I have also noticed that there is no entry in fstab for /dev/md0 although I can manually mount it using the same Disk Utility GUI. That is strange to me. Is it normal? i could easily add it manually but Ubuntu it won't boot anymore (i tried and failed, hence the reinstall). I tried for two weeks to find a solution browsing on different forums but the problem is beyond my expertise...
BELOW are further details about my configuration mdadm.conf, fstab, fdisk -l result and other info. I don=t want to loose my data but it would be nice to make this thing work and be able to access my fileserver via vnc instead of having to keep it connected to a lcd monitor as now. This is the blkid result:
I'm trying to find out which one is safer when it comes down to recovery process in case of a drive failure
A RAID5 created in mdadm or a Stripe RAID created on pure LVM
the RAID is purely for data storage for a SAMBA server, the OS will reside on its own drive.Ideally the RAID physical hard drives should be re-build on another machine in case of catastrophic server failure (mother board problem, or any other random problems as an example)I can't decide which of the 2 software RAID method is more convenient and safest, don't care about performance, it'll be a dedicated server for mass storage, it's going to mirror other 3 file servers on fakeRAIDs (dmraid), it's simply a redundant backup for the backups
The important goal here is portability.from what've read it appears that LVM might be more portable?but according to some dated (2009) info the mdadm seems to be a bit buggy when it comes to rebuilding the array, yet LVM doesn't appear that safe either which one would you pick for ease to rebuild on catastrophic failures?
I have an old Athlon XP 3000 machine that I keep around as a file server.It's currently got three 1TB drives which I had setup as mdadm raid 5 on FC10. The machine's original drive held the superblock for the raid array and it just had a massive heart attack. I've searched, my biggest source being URL...I can't tell if I can reassemble the superblock info lost with the original hard drive or if I've lost it all...
I am trying to build a new array after adjusting TLER on my disks, which permanently changed some of the drives sizes. I am not sure if the following inconsistencies are related to the newly mismatched drive sizes.
Code: mdadm --create --auto=md --verbose --chunk=64 --level=5 --raid-devices=4 /dev/md1 /dev/sdd /dev/sde /dev/sdf /dev/sdg Nets me (build-time was two full days):
On a side note, since I'm recreating my array from scratch, I was wondering if anyone here knows of any optimized settings I could use. I've got 3Tb of data to transfer, so lots of test material.
These are Western Digital First Generation 2TB Green Drives (WD20EADS-00R6B0) with WDidle3 fix applied & TLER=ON. These are pre Advanced Format (aka not 4K).
I have a 500GB drive mounted at /var/lib/mysql which is mostly empty and not part of any RAID array.The reshaping started and everything looked OK. The access lights on the 5 drives were all coming on at the same time on a regular basis. The status from /proc/mdstat showed the array being reshaped to RAID6, albeit slowly. The status showed an average speed of 4000KB and an estimated completion time of 4000 minutes. This all seemed reasonable. This was performed in late afternoon.
The next morning I checked the status and the average speed was down to 300->400KB and the estimated time to complete was 40,000 minutes. When I look at the drive lights, I have one drive whose access light is on solid and the other four drives come on intermittently. Running iotop doesn't show anything useful. mdadm and kjournal show up occasionally. The same is true for top (running on an i5 2500K Intel processor). Here is the output of cat /proc/mdstat:
I'm a bit sick to my stomach right now. I had a raid5 array (5x1.5TB drives) and I upgraded to lucid and now the array no longer works. Initially, on boot, it would try to mount it from fstab and that failed consistently as it wasn't assembling it.
then I tried to assemble it by hand (--scan) and that seemed to cause it to mount degraded (it seems md in the process tried to use on of the disks for something else!), but when I look at its partition table, it's empty. pretty pissed at the moment (somewhat at myself, didn't really need to upgrade), any ideas what went wrong?
This isn't exactly Ubuntu specific, but I do plan on using Ubuntu to try to recover this array. I've been using a Freedom9 freestor 4020 for the past few years and other than it totally blowing up last week it's been pretty good. I was on vacation for almost a month and a few days after I returned my NAS (freestor 4020) started acting up. I tried a few power cycles, but was dismayed to see that I could not log in via browser or SSH (SMB shares were no accessible either). A drive failure light is supposed to illuminate if a disk fails, but no dice.
I plugged all 4 drives from the NAS into an Ubutnu 9.04 Desktop system and one started throwing out all kinds of errors. Thinking that it would be a simple rebuild, I went to my local computer shop and picked up another 500GB drive (same manufacturer/part #), replaced the failed drive, and powered up the NAS again... Nothing. I left it for 12 hours then powered it down, plugged the new drive into my linux box again to see if it rebuilt... the drive was a virgin. What gives me hope that I can still recover the data is Ubuntu sees "RAID components" on the drives (through disk manager and parted), and gives me the option of initializing the array.
My plan of attack is to plug all of the drives back into my Ubuntu box, initialize the RAID array via LVM, and pretty much hope for the best. The data is not uber critical, but it would be a pretty big pain in the behind to rip/upload all the software that was on it (ripping hundreds of DVD/CD images is not fun). If my Ubuntu box can make sense of this newly initialized/mounted RAID set... I'll plug in a 2TB external drive, copy the data over, and rebuild the NAS from scratch, then put my data back on (perhaps a different unit, or something running openfiler).
I tryed to install ubuntu 10.04 using the beta alternative install cd.
Everything went fine until the partitioning section.
I choose manual partitioning and all my existing partitions were detected correctly included my 2 mdadm raid0 arrays.
I choose md0 as my / partition and choose to format the partition
I choose md1 as my /home partition as choose to keep the data
When I choose to continue and write the changes to disk the install started to create an ext4 partition on md0, the installer then stopped with an error that the kernel could not reread the partition table.
I aborted the installation at this point.
Now I can not access either of my arrays.
I have booted a livecd and installed mdadm. When I checked /etc/mdadm/mdadm.conf my existing arrays were already listed.
Code: # mdadm.conf # # Please refer to mdadm.conf(5) for information about this file. #
I had a raid array working great in 9.04 with mdadm, and I just recently upgraded to 10.04 (clean install) and I'm trying to reassemble the array and having a dickens of a time.When I try to recreate the array with:
Code: sudo mdadm --create /dev/md0 --level=5 --raid-devices=3 /dev/sdb /dev/sdd /dev/sdc I get this: