Ubuntu Servers :: Use UUIDs To Setup A Raid With Mdadm?

Oct 6, 2010

Can I use UUIDs to setup a raid with mdadm?

View 3 Replies


ADVERTISEMENT

Ubuntu Servers :: Create A New Mdadm RAID 5 Device /dev/md0 Across Three Disks?

Feb 5, 2011

I am trying to create a new mdadm RAID 5 device /dev/md0 across three disks where such an array previously existed, but whenever I do it never recovers properly and tells me that I have a faulty spare in my array. More-specific details below. I recently installed Ubuntu Server 10.10 on a new box with the intent of using it as a NAS sorta-thing. I have 3 HDDs (2 TB each) and was hoping to use most of the available disk space as a RAID5 mdadm device (which gives me a bit less than 4TB.)

I configured /dev/md0 during OS installation across three partitions on the three disks - /dev/sda5, /dev/sdb5 and /dev/sdc5, which are all identical sizes. The OS, swap partition etc. are all on /dev/sda. Everything worked fine, and I was able to format the device as ext4 and mount it. Good so far.

Then I thought I should simulate a failure before I started keeping important stuff on the RAID array - no point having RAID 5 if it doesn't provide some redundancy that I actually know how to use, right? So I unplugged one of my drives, booted up, and was able to mount the device in a degraded state; test data I had put on there was still fine. Great. My trouble began when I plugged the third drive back in and re-booted. I re-added the removed drive to /dev/md0 and recovery began; things would look something like this:

Code:
user@guybrush:~$ cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdc5[3] sdb5[1] sda5[0]
3779096448 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]

[Code]...

View 9 Replies View Related

Ubuntu Servers :: Safely Remove Mdadm RAID Superblock

Apr 5, 2011

I bought a disk to a friend that used it in a raid array, using the entire disk for the raid usage. To put that disk on service, i used dd-rescue to copy my old disk entirely, and managed to grow and setup a the partition table without losing any data. My last step was to create a RAID between my entire old disk, with a single partition and a partition of the same size on my new disk. I ran into some problems, but i manage to somehow fix it imperfectly, but now this setup is working properly. The problems (and imperfection) came from an issue it did not suspected : at some point, the original RAID superblock of the new disk, living in /dev/sda, resisted to dd-rescue, and so it is scanned by mdadm that tries, obviously unsuccessfully, to use it.

Partition layout :

Code:

Disk /dev/sda: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

[code]....

this setup is working properly besides this raid5 declared on sda, so that is shows up here and there. Since it is using the same device name that my other, proper raid setup, i don't know how to deactivate it since mdadm uses the /dev/mdx name to identify arrays.

View 4 Replies View Related

Ubuntu Servers :: Mdadm RAID 6 Array With Si 3132 SATA Controller ?

Mar 12, 2010

I've recently started having an issue with an mdadm RAID 6 array that been operational for about 2500 hours.

Intermittently during write operations the array stalls, dropping to almost 0 write speed for 10-30 seconds. When this occur one or both of the 2 drives attached to a 2 port Silicon Image si3132 SATA-II controller "locks up" with its activity light locked on. This just started occurring within the last week and didn't seem to coincide with any update that i noticed. The array has just recently passed 12.5% full. The size of the write does not seem to make any difference and it seems completely random. Some times copying a 5 GB dataset results in no slow down other times a torrent downloading to the array at 50kb/sec does cause a slow down and vise versa.

The array consists of 8 WD 1.5TB drives, 6 attached to the ICH9R south bridge, and 2 attached to a si3132 based PCI express card. The array is formatted as a single ext4 partition.

Checking SMART data for all drives shows no errors. Testing read speed with hdparm reports what i would expect (100mb/sec for each drive, ~425mb/sec for the array).

The only thing i did notice is that udma6 is enabled for all the ICH9R drives while only udma5 is enabled for the si3132 drives. Write cache is enabled for all the disks. Attempting to set the si3132 drive to udma6 results in an IO error from hdparm.

The si3132 drive is using the sata_sil24 driver. Nothing of interest appears in the kern or syslog. During this time top shows very high wait time.

The s13132 controller appears to have the original firmware from 2006 loaded, there are some firmware updates available on the Silicon Image website for this controller that now appear to offer separate firmwares for RAID operation (some sort of hybrid controller/software thing the controller supports) and a separate firmware for standard IDE use.

Has anyone had similar issues with this controller? Is a firmware update a reasonable course of action? If so which firmware is best supported by the linux driver?

I know i'm not using its raid features but i've dealt with controllers that needed to be in raid mode for ahci to be active and for linux to work well with them. I'm bit ify at the idea of just trying it and finding out as it could knock 2 disks of my array out of action.

View 2 Replies View Related

Ubuntu Servers :: Mdadm - Why /dev/sdb1 And /dev/sdi1 Show As Both Ext2fs And Also As Part Of A RAID Array

May 31, 2011

I've been having some problems w/ a my RAID 5 array, and after extensive investigation, I'm fairly sure that my last resort is rebuilding the array. I'd tried --assemble, b/c it's a previously created array, but it didn't seem to like that. So, I checked into --create, and it will re-create the array w/out destroying the data, if the superblocks are persistent, which they seem to be. However, here's what I get:

[Code]....

My question is: why do /dev/sdb1 and /dev/sdi1 show as both ext2fs and also as part of a RAID array?

View 3 Replies View Related

Ubuntu :: Mdadm Raid + GRUB = Not Booting - Error: Unsupported RAID Version: 0.91

Jul 18, 2011

I have a raid5 on 10 disk, 750gb and it have worked fine with grub for a long time with ubuntu 10.04 lts. A couple of days ago I added a disk to the raid, growd it and then resized it.. BUT, I started the resize-process on a terminal on another computer, and after some time my girlfriend powered down that computer!
So the resize process cancelled in the middle and i couldn't acess any of the HDDs so I rebooted the server.

Now the problem, the system is not booting up, simple black with a blinking line. Used a rescue CD to boot it up, finised the resize-process and the raid seems to be working fine so I tried to boot normal again. Same problem. Rescue cd, updated grub, got several errors: error: unsupported RAID version: 0.91. I have tried to purge grub, grub-pc, grub commmon, removed /boot/grub and installed grub again. Same problem.

I have tried to erased mbr (# dd if=/dev/null of=/dev/sdX bs=446 count=1) on sda (ide disk, system), sdb (sata, new raid disk). Same problem. Removed and reinstalled ubuntu 11.04 and is now getting error: no such device: (hdd id). Again tried to reinstall grub on both sda and sdb, no luck. update-grub is still generating error about raid id 0.91 and is back on a blinking line on normal boot. When you'r resizeing a raid MDADM changed the ID from 0.90 to 0.91 to prevent something that happend happened. But since I have completed the resize-process MDADM have indeed changed the ID back to 0.90 on all disks.

I have also tried to follow a howto on a similar problem with a patch on [URL] But I cant compile, various error about dpkg. So my problem is, I cant get grub to work. It just gives me a blinking line and unsupported RAID version: 0.91.

View 2 Replies View Related

Ubuntu Servers :: Custom RAID Setup ?

Jul 12, 2010

I'm setting up a web server but I have no experience with RAID. I would like to try this configuration if possible:

2 x HDD 500GB RAID1
1 x HDD 20GB (logs and tmp)

The old 20GB drive I would like to use it to store logs and temporally files (mounted in /var/log and /tmp respectively). With this I'm trying to reduce some disk usage in the RAID drives. In my idea, it would be better to write the access/error logs of the web server in a separated drive to the one serving the files which may increase speed... sounds crazy?

One problem is that during the installation, If I set the RAID automatically it will try to use my 20GB HDD as well in the RAID... Does it will work if I set the RAID first (removing the 20GB HDD) and then set the mount points in it after the installation?

View 4 Replies View Related

Ubuntu Servers :: Init Not Found After RAID Setup?

Apr 4, 2011

I have the current disk setup :

disk1 :

Code:
sda:
(primary) sda1 ntfs
(primary) sda2 ext3 (mounted as / )[code]....

Before i set up the raid, but with this exact partitioning, the system booted perfectly. When i installed mdadm and created the raid1 mirroring on sda6 and sdb1, the init got screwed up, and all i get is a shell on initramfs, from where i can inspect that sda is binded on md, and cat /proc/mdstat tells me that i have an inactive sda[4].I can't mount the root partition (sda2), because it's busy (i suspect dmraid to lock it), which is, i guess, why init cannot be found.

I wonder if my error is to setup a raid array using a logical partition contained in an extended partition (but i hardly see why it would not work - but the sda bind and the sda[4] in mdstat seems to tell me that it does not), or it's just the initrd that is improperly configured. The other things that bothers me, is that changing the partition type of the raid partitions (fd to 0 - Empty), to disable raid autodetection, resulted in the same behavior on boot. Which might lead me again to think about configuration file problem instead of improper setup.The live cd doesn't not seem to recognize raid, so i can't inspect problems any further, but i could inspect system configuration, but i don't really know where to start.

View 2 Replies View Related

Server :: Setup The Servers In A RAID Configuration?

Apr 15, 2010

I am rebuilding two microsystems servers and I need some advice to make my dreams come true.I want to setup the servers in a RAID configuration and want to install a GUI Linux application to manage a file server, manage a subnet, and host a Moodle on my subnet.I am planning to use Asus eee netbooks running Linux as my client computers. I basically need to be able to get my kids on the web and be able to have them use some open source office suite tools. No major crunching. I'll have two Macs for that.

View 3 Replies View Related

Ubuntu Servers :: Install Boot Loader Into RAID 10 Setup?

Oct 3, 2010

i installed ubuntu server on a 4-hard disk system, i installed with the RAID 10 support that comes with 10.04 server optioon. my raid swap is /dev/md0 and my raid data partition is /dev/md1

hard drives are /dev/sda /dev/sdb /dev/sdc /dev/sdd

the instalation goes fine, from the 8TB im getting 4TB mirrored which is what i want,
but the instalation setup goes up to the point where it asks me to install grub boot loader, There i choose the default option and it does a fatal error.

my main goal is just to make it bootable, any method is welcome. this went beyond my RAID knowledge (more conceptual than technical)

View 2 Replies View Related

Ubuntu Installation :: Migrate Working Single Disk System To Existing RAID Array Using Disk UUIDs

Aug 1, 2010

I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.

I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).

These are the commands I used:

Quote:

p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*

[Code]....

I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...

So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.

It's still trying to use /dev/disk/by-uuid/412d... to boot.

What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.

View 2 Replies View Related

Ubuntu Servers :: Not Quite Understanding UUIDs And Grub2 On 10.04 Server?

Jul 16, 2010

I'm trying to set up an Ubuntu server using 10.04 (64-bit), and running into problems after a couple of reconfigurations. Here's the full story:I initially built the server on a 400+GB RAID5 array, putting everything but swap in one partition. Unfortunately, I needed to repartition, putting / in the first primary partition, swap in the second partition, /var/log in the third primary partition and /home in the remaining space on the fourth primary partition.

However, at this step, I ran into some problems with UUIDs in /etc/fstab and Grub2 (I've used Linux for about 9 years, but I'm new to Ubuntu, and I haven't used UUIDs or Grub2 on Gentoo, yet). Consequently, I made the (probably not smart) decision to move back to the /dev/sdX notation I am familiar with.The problem with that is that now I need even more space on /home, so I've added a Dell Powervault and a Dell PERC5/e SATA card to my server. Now, Grub2 tries to boot from the new RAID array on the Powervault instead of the internal RAID array, so I am trying to move back to the UUID notation so that I don't have worry about /dev/sda being the internal array sometimes and the external array at other times.I don't mind being RTFM'd, but I'm having trouble finding pointers to the documentation explaining Grub2 configuration and the UUID notation. Does anyone have pointers to some readable, concise documentation on configuring this in Ubuntu?

View 4 Replies View Related

Fedora Servers :: Setup RAID On DELL Server With F10 64Bit OS?

Jan 22, 2009

I have had no problem installing Fedora OS on any of my Dell servers prior to this post. Anyway, I wouldn't call this a problem but recently, we bought another DELL server with Quad Core, 4GB, etc... AND this model has 2 swappable SAS Harddrives.

I wouldn't call myself an expert but then again I am not a newbie too. However, I have never setup any RAID before and now I am forced to setup RAID1 on this server. So, in a way, I am a newbie in setting up RAID

Would someone please point me in the right direction as I have no idea what I am supposed to do to setup the RAID. FYI, I will be installing Fedora 10 64bit on this server. I would appreciate if you can start from the very beginning, ie. partitioning, formatting the harddrives during OS installation, etc..

View 3 Replies View Related

Fedora Servers :: RAID 6-like Setup With Dissimilar Disk Sizes?

Jun 7, 2010

I'm looking to set up a bit of a home server, and am wondering about storage. What I'd like is something like RAID 6, which has good redundancy built-in, but with this being a home server, I'd prefer to start a little smaller and leave room to build it up in future. I'd been looking at commercial products like the 'drobo', which seems fairly ideal, but I'd really like to see if I can do it myself. I understand that throwing the RAID into an LVM will allow for some expansion, but the last time I checked, most RAID setups called for the same sized disks, or at least limited the array by the size of the smallest disk present.

What I'd like is the ability to build a basic framework with a few cheap disks, and then as things start filling up, to be able to add larger ones (perhaps eventually pulling out smaller ones as though they'd failed and replacing them with big ones)

View 1 Replies View Related

Ubuntu Servers :: 9.10 Server/Raid 10 Setup - HDD - Not Recognizing The Full Capacity Of The Drives

Mar 12, 2010

I have a total of 4 hdd's, 500gb 7.2rpm that I would like mirrored using raid 10. As you can see from the image, ubuntu 9.10 server isn't recognizing the full 2tb's. In fact, I'm not even sure about the configuration as I was thinking the HDD's would come up as four 500gb hdds. Instead I have the configuration above set and ready for Ubuntu to be installed on.

1. Is this typical of a server pre-configured from Dell(perc6 raid controller.

2.Why is ubuntu not recognizing the full capacity of the drives especially when it's a server install?

View 5 Replies View Related

Ubuntu Servers :: Tell Grub To Stop Trying To Boot Based On UUIDs?

Dec 16, 2010

how do I tell grub to stop trying to boot based on UUIDs? I've been dealing with this for about two years now. I have this one system, a PowerEdge 1550, that for some reason WILL NOT boot using a UUID, and will only boot if I edit the first entry in grub, remove everything to do with UUIDs (completely delete the search line, replace root=uuid=[uuid] with root=/dev/sda1). I can't seem to make these settings stick and now with 10.10 there's no longer a menu.lst to edit. So, in short, I need grub to stop trying to boot based on a UUID. Absolute paths, baby. That's what I need.

View 3 Replies View Related

Ubuntu :: MDADM Raid 5 - OS Failure?

Jun 5, 2011

I have 4 WD10EARS drives running in a RAID 5 array using MDADM.Yesterday my OS Drive failed. I have replaced this and installed a fresh copy of Ubuntu 11.04 on it.I then installed MDADM, and rebooted the machine, hoping that it would automatically rebuild the array.It hasnt, when i look at the array using Disk Utility, it says that the array is not running. If i try to start the array it says : Error assembling array: mdadm exited with exit code 1: mdadm: failed to RUN_ARRAY /dev/md0: Input/output errormdadm: Notenough devices to start the array.I have tried MDADM --assemble --scan and it gives this output:mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.I know that there are 4 drives present as they are all showing, but it is only using 2 of them.I also ran MDADM -- detail /dev.md0 which gave:

root@warren-P5K-E:~# mdadm --detail /dev/md0
/dev/md0:
Version : 0.90

[code]....

View 2 Replies View Related

Ubuntu Servers :: RAID-5 Recovery (spare/active) / Degraded And Can't Create Raid ,auto Stop Raid [md1]?

Feb 1, 2011

Could any RAID gurus kindly assist me on the following RAID-5 issue?I have an mdadm-created RAID5 array consisting of 4 discs. One of the discs was dropping out, so I decided to replace it. Somehow, this went terribly wrong and I succeeded in marking two of the drives as faulty, and the re-adding them as spare.

Now the array is (logically) no longer able to start:

mdadm: Not enough devices to start the array.Degraded and can't create RAID ,auto stop RAID [md1]

I was able to examine the disks though:

Code:
root@127.0.0.1:/etc# mdadm --examine /dev/sdb2
/dev/sdb2:
Magic : a92b4efc
Version : 00.90.00
code....

Code:
mdadm --create --assume-clean --level=5 --raid-devices=4 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2
As I don't want to ruin the maybe small chance I have left to rescue my data, I would like to hear the input of this wise community.

View 4 Replies View Related

General :: Possible Mdadm RAID States?

Jul 11, 2010

I'm writing a monitoring plugin for a home server RAID, mdadm on Ubuntu 10.4. code...

I'm looking for the possible values of "state" but can't seem to find it anywhere, neither man nor the online documentation I have found seem to have a list.

Does anyone know where to find a list of possible states?

View 1 Replies View Related

Software :: RAID Created With 'mdadm' On /dev/md0, But Uses /dev/md3?

Mar 10, 2010

I have done lots of searching and I haven't been able to find anyone else with the same problem. Whenever I create a RAID with 'mdadm', regardless of level (I've done linear, 0, and 5) the command I use is:

Code:
mdadm --create --run --verbose /dev/md0 --raid-devices=11 --spare-devices=1 --chunk=256 --level=5 /dev/sd[abcdefghijkl]1

The RAID is build RAID 5 as it should be. However, when I check /proc/partitions it shows under "md3".

[Code]...

View 1 Replies View Related

CentOS 5 :: Mdadm Not Assembling Raid?

Oct 21, 2010

I have a previously defined RAID5 (4 disks). This worked in Ubuntu 8.04.. I recently moved to CentOS5.. and I cannot seen to get the drive back online.cat /proc/mdstat shows, no raid levels (personalities).. and no drives listed.mdadm --detail -scan returns nothing.mdadm --QE returned a UUID string.. and the ARRAY output.I can mdadm --examine all the members of the original array.I am not versed in mdadm enough to really understand what I can run and should not run that would erase the data on the drives. Please assist.. I will try to post exact output of commands.. but the system is kind of unreachable and being rebuilt... i just want to ensure my data on the array is not lost

View 7 Replies View Related

Ubuntu :: Mdadm Raid 5 Single Disk Failure

Feb 2, 2010

Recently, one the SMART utility said that one of the drives had failed and another drive was about to fail. I downed the box and hooked them up to my windows machine to run sea tools on them (They are all seagate drives). Sea Tools said that the drives were fine, while ubuntu said they were failing/dead. Yesterday I decided to try to fix one of the drives in the raid. I turned the server off, took the failed drive out, and restarted. Of course the raid didn't work because only 2 of the 3 drives were there, however it had been working w/ only 2 of the 3 drives for a couple months now (I'm a lazy college student). I turned it back off and back on with the drive there just to see if I could get the raid up again, but I havn't been able to get it to go. So far I've tried:

Code:

mdadm --assemble /dev/md0 /dev/sd[b,c,d]
mdadm: no recogniseable superblock on /dev/sdb
mdadm: /dev/sdb has no superblock - assembly aborted

[code]....

I'm looking for a way to trick the raid into working with just 2 drives until I can warranty the seagate and buy an external 1.5 TB drive to use as another backup. how to remove the bad drive from the array and replace it with a fresh drive, without data loss.

View 3 Replies View Related

Ubuntu :: MDADM RAID 5 Failed But Disks Are Still Present?

Jun 7, 2010

I just had a whole 2TB Software RAID 5 blow up on me. I rebooted my server, which i hardly ever do and low and behold i loose one of my raid 5 sets. It seems like two of the disks are not showing up properly.. What i mean by that is the OS picks up the disks, but it doesnt see the partitions.

I ran smartct -l on all the drives in question and they're all in good working order.

Is there some sort of repair tool i can use to scan the busted drives (since they're available) to fix any possible errors that might be present.

Here is what the "good" drive looks like when i use sfdisk:

Quote:

sudo sfdisk -l /dev/sda
Disk /dev/sda: 121601 cylinders, 255 heads, 63 sectors/track
Units = cylinders of 8225280 bytes, blocks of 1024 bytes, counting from 0
Device Boot Start End #cyls #blocks Id System
/dev/sda1 0+ 121600 121601- 976760001 83 Linux
/dev/sda2 0 - 0 0 0 Empty

[Code]....

View 2 Replies View Related

Ubuntu :: MDADM RAID 5 Disk Failure And Recovery?

Jun 18, 2010

I have a fileserver which is running Ubuntu Server 6.10. I had a RAID5 array consisting of the following disks:

Code:
/dev/sda1
/dev/sdb1
/dev/sdd1
/dev/md0 -

the raid drive for the above three disks. The sda1 disk has failed and the array is running on 2 of 3 disks

/dev/sdc (OS disk)
/dev/sde (new 2tb disk - unused)
/dev/sdf (new 2tb disk - unused)

My plan was to rebuild the array using the two new disks as RAID1. Would the best way to do this be to create a new RAID1 disk on /dev/md1 then copy all data over from /dev/md0? Also - this may sound stupid but since all 3 drives in md0 are identical i'm not sure physically which disk is bad. I tried disconnecting each disk one-by-one then rebooting but the system doesn't appear to want to boot without the bad drive connected. I've already failed the disk in the array with mdadm but i'm unsure of how to remove it properly.

View 3 Replies View Related

Ubuntu :: Server Was Used To Boot (after MDADM Raid Rebuild)

Feb 1, 2011

My home-backup server, with 8*2TB disks won't boot anymore. Two disks failed at the same time and i rebuilt the raid 6 array without any problem, but now i can't boot the os. I'm using ubuntu server, 10.10. I've made screens of the displays to don't copy everything here. The problem at the boot:

And the Grub config: It's not a production server, but i would like to have it online. I've tried for the lasts 2 days (just a couple hours a day) but without success. I was suggested to do "mount -o remount,rw /" and than edit /etc/fstab, but it get the file don't exist error.

View 2 Replies View Related

Ubuntu :: 10.10 Mdadm RAID - Moving Device Names ?

Mar 25, 2011

I'm having trouble with Ubuntu 10.10 and stable device names. When I installed Ubuntu, the root drive was the only one in the machine; it obviously got /dev/sda.

After the base installation, I installed three additional 2TB drives to make RAID-5 array. Ubuntu renamed the root drive to /dev/sdd. While annoying I lived with it.

After creating a single partition set to "Linux raid autodetect" on each drive, I created the RAID-5 array:

Code:

All was going well until a reboot. When rebooting Ubuntu decided to make the root drive /dev/sda this time and now mdadm --detail /dev/md0 reports:

Code:

How to fix the array and make the device names stable?

View 1 Replies View Related

General :: Bad Sectors On Mdadm Raid 5 Array?

Aug 14, 2010

I'm running a Debian homeserver, with a 3-disk (1GB each) raid 5 array using mdadm (the OS is on a separate disk).Now, smartmontools noticed some bad sectors on one of the disks, and I'm not sure what to do next (except for backup of valuable data).I found some articles on how to fix these sectors, but I'm unaware what the result on the whole array will be.

View 4 Replies View Related

Server :: Weird With Software Raid-5 Mdadm?

Feb 20, 2011

My only goal is to have a raid-5 that auto-assembles and auto-mounts. Hardware: 4*2TB sata (raid disks), 1*500GB IDE (OS disk), 1*DVD IDE all plugged direct into the motherboard (nForce 750i SLI).

Starting partitions on the raid disks: gpt ext4 The problem occurs when I restart my comp after building it for the first time. I am able to see it assemble, I am able to partition it, I even mounted it Once.This is the second time I've built it so I have watched everything that happened. I don't know if this has anything to do with my problem, but when I created the raid my drive designations were: sda - 500GB(OS), sd[bcde] - 2TB(raid). When I restarted: sd[abcd] - 2TB(raid), sde - 500GB(OS).

[Code]...

View 3 Replies View Related

Software :: Mdadm Is Letting Me Down - Won't Recover My Raid

Aug 21, 2010

I HAD a fedora 11 server with md RAID 1 across two 1TB SATA drives. The md0 space was set up to be an LVM PV and the single LVM VG was carved up into 5 or 6 LVs. The MB on this system died and I wound up buying a new one.

Now I want to recover the data from the RAID1 setup on the new server. However, when I attach the two 1TB drives to a new fedora 13 setup, mdadm is only able to find one of the two drives. The partition on the second drive shows "busy" during an mdadm -A -s -v to scan for md volumes.

Well, one drive should be enough since this is RAID1, right? Well, when I do a pvscan -v, the other drive shows up as a "NEW" pv not allocated to a VG. In addition, vgscan does print "Invalid metadata header checksum" when it runs but it doesn't point at any particular PV. I'm afraid to go any further with LVM since I can't afford to lose the data on this system. It is backed up offsite, but the restore will take several days and I can't afford to be down that long.

Are there any tools or techniques where I can dig deeper into what each drive, in the RAID1 pair, has right and wrong with it and pick one that I can force into a usable VG so that I can recover the data?

View 2 Replies View Related

Software :: Creating New Mdadm Raid 1 Array?

Mar 2, 2011

a server that was running a hardware isw raid on the system (root) disk. This was working just fine until I started getting sector errors on one of the disks. So, I shutdown the system and removed the failing drive and installed a new drive (same size). On reboot I went in to the intel raid setup and it did show the new drive and I was able to set it to rebuild the raid. So, continuing the reboot everything came up just fine except the raid 1 on the system disk. I have tried many times to get the system to rebuild the raid using dmraid, but to no avail it would not start a rebuild. In order to get the system back up and make sure that the disk was duplicated I was able to 'dd' the working disk to the new disk that was installed.At present when I look at the system it does not show up with a raid setup on the system disk ( this comprises the entire 1TB disk with w partitions sda1 as / and sda2 as swap).Problem:I have decided to forego the intel raid and just use mdadm. I have a test system setup to duplicate (not the software, but the disk partitions) the server setup.

Code:
[root@kilchis etc]# fdisk -l
Disk /dev/sda: 1000.2 GB, 1000204886016 bytes
255 heads, 63 sectors/track, 121601 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

View 12 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved