General :: No Automatic Rebuild Of RAID 5 After Replacing Bad Disk

Dec 12, 2009

I have a 5 disk raid 5 array that is composed of SATA A:0,1; SATA B: 0,1, and SATA C:0, and one of the disks (SATA A:0) recently went bad on me. I have an ICP raid controller that is about 5 years old. I replaced SATA A:0. After rebooting, I went into the controller and verified that it saw the disk in the hard-disk info section...there I noticed that in the "status" section, that the SATA C:0, SATA B:1 disks were listed as being "in array", the SATA A disks were blank, and the SATA B:0 disk was listed as "fragment". When I go into the "repair array" section, the controller tells me that there are no arrays that are in failure, error, or need to be rebuilt.

This puzzles me, as I thought the controller would know that the array needs to be rebuilt after replacing the disk and I don't see a way to initiate a rebuild. If I just let the server boot after replacing the disk, then I get back that there are the correct number of disks in the raid 5 and that it is ready, however, the screen then goes blank and I get a blinking cursor and the system seems to hang. There are no activity lights on any of the drives associated with the raid 5, which makes me think that the system is not rebuilding the array at this point.

View 3 Replies


Ubuntu :: Rebuild Md RAID Array After OS Disk Failure?

Dec 19, 2010

I went to setup my linux box and found that the OS drive had finally died. It was an extremely old WD raptor drive in a hot box full of drives so it was really only a matter of time before it just quit on me. Normally this wouldn't be such a big deal however I had just recently constructed an md RAID5 array of 3 1TB disks to act as an NFS mount for basically all of my important files. Maybe 2-3 weeks before the failure I had finished moving all of my most important stuff onto that array. Now I know that the array is intact. All the required data is sitting on those disks. Since only the OS level disk failed on me I should be able to get a new disk in there, reinstall ubuntu and then rebuild that array. how exactly do I go about doing that with mdadm? Do I create the array from the /dev character devices like when I initially built the array?

View 2 Replies View Related

Ubuntu Servers :: Does Replacing Failed Disk On Raid5 Change Any Data On Other Disks In Raid

Jan 9, 2011

I've got a raid5 array of 4 disks with ubuntu 8.04 runing on it that is currently still working:


Smartmontools for /dev/sdc tell that there are 9 sectors pending for reallocation:


197 Current_Pending_Sector 0x0012 100 100 000 Old_age Always - 9
198 Offline_Uncorrectable 0x0010 100 100 000 Old_age Offline - 9
And /dev/sdd has increasing number of reallocated sectors (about 1 every couple of minutes):


5 Reallocated_Sector_Ct 0x0033 100 100 036 Pre-fail Always - 1735
/dev/sdc has failed a coulple of times this week (but I have always sucessfully readded it to raid5) . But the increasing number of reallocated sectores on /dev/sdd concerns me even more.

I'm affraid that during removal of /dev/sdd and adding new /devs/sdd disk, raid might fall appart. That's why I would try to do it in Ubuntu Live CD:If the raid falls appart (/dev/sdc fails) during the readding of new /dev/sdd disk, I might still remove the new /dev/sdd and return the previous one and assemble the raid with:

/dev/sdd (old one that was previously removed)

Does assembling Raid in Ubuntu Live and adding new disk for /dev/sdd write anything on /dev/sda, /dev/sdb and /dev/sdc in the process of adding /dev/sdd into raid5?

View 2 Replies View Related

CentOS 5 :: How To Rebuild Sofware Raid

Apr 6, 2009

I've got a server running a software Raid for SATA disks on a P5E motherboard.

I had to had a lot of memory on thi sserver, then I have had to flash the bios. This has resetted the software raid on the disks, then when I boot, i got a Kernel Panic cause it doesn't find anything ...

how to rebuild the raid ? I can boot on a live cd, or anything else, but don't know how to do it without loosing my data.

View 2 Replies View Related

Ubuntu :: Server Was Used To Boot (after MDADM Raid Rebuild)

Feb 1, 2011

My home-backup server, with 8*2TB disks won't boot anymore. Two disks failed at the same time and i rebuilt the raid 6 array without any problem, but now i can't boot the os. I'm using ubuntu server, 10.10. I've made screens of the displays to don't copy everything here. The problem at the boot:

And the Grub config: It's not a production server, but i would like to have it online. I've tried for the lasts 2 days (just a couple hours a day) but without success. I was suggested to do "mount -o remount,rw /" and than edit /etc/fstab, but it get the file don't exist error.

View 2 Replies View Related

Fedora :: Eth0,1,2,3 Become Eth4,5,6,7 After Cloning Via Raid 1 Rebuild?

Aug 10, 2011

I've cloned a machine by removing a HDD from a Raid mirror set putting it into another machine and powering on and fsck'ing a few times and everything is great apart from the ethernet port numbering has gone a bit wonky.

The cloned machine which must have a specific configuration using eth0,1,2,3 naming convention however when I boot the freshly cloned machine eth4,5,6,7 show up and eth0,1,2,3, are missing (as it detects the new machines ethernet ports)

Is there some configuration file I can delete, reboot the machine and then Fedora rebuilds the file using eth0,1,2,3, and populating with the correct hardware address.I need to clone this machine lots and lots and lots of times and the manual why I've figured out is a little long winded.

View 2 Replies View Related

CentOS 5 Server :: Random Auto Rebuild Of Software Raid ?

Jan 16, 2010

I have set up a software raid 5 and a software raid 1 on the same 3 disks.

But my problem is, every couple of days the software raid starts automatically resyncing. This resync is always succesful.

I do not know why this happens, any ideas to troubleshoot this further? I can not see an error in /var/log/messages

View 14 Replies View Related

Ubuntu Servers :: RAID Arrays Rebuild The Data On The New Drive?

Jun 5, 2010

I have never preformed a rebuild of an RAID array. I am collecting resources, which details how to build an RAID 5 array when one drive has failed. Does the BIOS on the RAID controller card start to rebuild the data on the new drive once it is installed?

View 4 Replies View Related

Server :: Mdadm: Rebuild A Software Raid From Drives With Existing Partitions?

Apr 24, 2011

Its from a Synology Box with 3 disks, which one is damaged. But this disk wasnt in use.Take a look on the raid-size of 493 GB - and the both available disks with 250GB..)
On the others there were a linear raid. during this damaged disk the synology-device tells me, that the volume was crashed.But it look like, that this disk was not mounted into this volume.Quote:

DiskStation> mdadm --detail /dev/md2
Version : 00.90


View 3 Replies View Related

Ubuntu Servers :: Resize Raid Partition - Rebuild /dev/md2 Without Lost Any Datas ?

Jun 15, 2010

I've got 2 hard disk drives of 250GB, with this partitions :

All partitions are with format "linux raid software". I made 3 raid1 with them like that :

I want to clone them to news hdd of 500Gb, to have this partitions :

My problem, how can I resize partitions, before resize the array md2 with the command mdadm --grow /dev/md1 --size=max ?

After, how can I rebuild /dev/md2 without lost any datas ?

I'm using ubuntu-server 10.04 64bits.

View 3 Replies View Related

Server :: Mdadm: Avoid Complete Rebuild Of RAID 6 Array (6/8 Active Devices)?

Jun 25, 2008

OS: Ubuntu 7.10 i386 Server (2.6.22-14-server)upgraded toUbuntu 8.04 i386 Server (2.6.24-19-server)I have 8 SATA drives connected and the drives are organized into three md RAID arrays as follows:/dev/md1: ext3 partition mounted as /boot, composed of 8 members (RAID 1) (sda1/b1/c1/d1/e1/f1/g1/h1)/dev/md2: ext3 partition mounted as /root, composed of 8 members (RAID 1) (sda2/b2/c2/d2/e2/f2/g2/h2)/dev/md3: ext3 partition mounted as /mnt/raid-md3, composed of 8 members (RAID 6) (sda3/b3/c3/d3/e3/f3/g3/h3), this is the main data partition holding 2.7TiBs worth of dataAll the raid member partitions are set to type "fd" (Linux RAID Autodetect).Important Note: 6 of the drives are connected to two Sil3114 SATA controller cards whilst 2 of the drives are connected to the on-board SATA controller (I don't know which model it is).

After upgrading my Ubuntu installation to 8.04, upon system restart there was an error message saying that my RAID arrays were degraded and thus the system was unable to boot from it.At the time, not knowing the cause of the sudden RAID failure, I attempted to force mdadm to start the arrays anyways (the RAID 1 arrays with 8 members each were no causes for concern, of course, but I wanted to back up my data on the degraded md3 array as soon as possible).Then it hit me, why would it recognize only 6 drives? Apparently the kernel has some compatibility problems with certain SATA controllers and my on-board controller chip was one of them.Sure enough, after moving all 8 drives to the Silicon Image controllers, the drives were all recognized without any problems.

If the missing drives were recognized again before the array was ever brought up again, everything would've been fine. But unfortunately I forced mdadm (--run switch) to bring it online with 2 missing members.This is when the problem began. I know that as soon as I re-add the two missing drives back into the md3 (RAID 6) array, the system will attempt to rebuild the array, using the data from the 6 drives.Given the size of the array and the type of the disk drives being used (off-the-shelf SATA drives with bit error rate of 1 out of 10^14 bits), I think it is highly likely that the system will encounter one or more bit errors during the rebuild., at this stage what I'm wondering is:1. If mdadm encounters a bit error during a RAID 6 rebuild, will it just give up on that particular file and move on to recover other data on the array? Or will it trash the entire array?

2. Is it possible to cheat mdadm by somehow replacing the new "raid metadata" on the 6 drives with the old data on the 2 drives? Will it make mdadm think the array is clean, consistent and nothing ever happened? Please do note that I did not write ANY new data onto the RAID 6 array from the time it was degraded until the time I brought it down with (--stop).Sorry for the long post and thank you for your time in advance. I really hope to get this RAID array back up without data corruption because I don't have a working backup of the array (I know, very stupid of me

View 4 Replies View Related

Ubuntu :: Replacing With Larger Drives In RAID 6?

Jun 14, 2011

I have a RAID 6 built on 6x 250GB HDDs w/EXT4. I will be upgrading the RAID to 4 2TB HDDs.

How would one go about this? What commands would need to be ran? I'm thinking about replacing the drives 1 at a time and letting it do the rebuild, but I know that would take a lot of time (which is fine). I don't have enough SATA ports to setup the new RAID and copy things over.

View 3 Replies View Related

Fedora :: Replacing Dead Disk In Lvm?

Jul 4, 2010

I know that this topic has been posted, responded to, and maybe even resolved, many times here, but I am stuck here with partially dead fileserver and need some pointers.

Problem: one disk drive that was part of a logical volume died. I have a replacement, but I can't get it into the LV and get the LV back up again.

pvcreate --uuid <uuid of dead drive> /dev/sdX1, where /dev/sdX1 is the newly created drive and its partition.
vgcfgrestore VolGroup
vgscan VolGroup
vgchange -ay VolGroup
e2fsck /dev/mapper/VolGroup-LogVol

but, e2fsck can't find a superblock. Apparently this drive is the first in the LV sequence, and it is not formatted as part of the LV.

So how to I get this new disk formatted into the LV without reformatting the entire LV and losing what data I still have?

(FWIW: this is on F11)

View 2 Replies View Related

Ubuntu Installation :: Rebuild Status Raid 1 Intel Isw Fakeraid Intel?

Oct 1, 2010

I have recently installed Ubuntu 10.04.1 lts server on my Intel "fakeraid" (software raid) (2x250 sata).To test my RAID 1 I turned off one HD and start the system.The first screen (Intel software screen) show Status = Degraded, but the system starts normally with just one HD.Then I turned off the system and turned on the HD again, so the first screen (Intel software screen) shows Status = Rebuild. If I enter in the software raid panel the folowing message is showed: "Volumes with "Rebuild" status will be rebuilt within the operating system"The system starts normally... but this message status stays permanently even I restart the system again

View 9 Replies View Related

Software :: Rebuild/repair Array(Raid 1) With Only "mdadm" Command Slack12.2?

Mar 18, 2010

I wonder how to attach new sata hard disk to software array where are two disk and one is crashed (this is a mirroring mode=Raid 1).Situation like this:I unpluged crashed disk and I buy the similar one and plug in What Next should I do?

View 4 Replies View Related

CentOS 5 :: Software RAID Auto Sensing Auto Rebuild Possible?

Jul 23, 2010

I recently installed a server with Software RAID. I tested by powering it down, unplugging one drive and powering it up. Magically, it worked!I found out later that I have to manually add individual devices like md1 to sda2 md2 to sda4. I got all of them added and rebuilt but my question is: Is there a way to make it so that if I "removed" a drive and put it back, the system will senses the new drive and rebuilds based on some internal table?

View 1 Replies View Related

General :: Raid - RAID1 With Only One Disk

Oct 27, 2010

I have a hypothetical situation in which I installed my operating system using a RAID1 mirror. At some point I decided that this setup was overkill, my machine isn't system critical, I value doubling my storage space more than speedy recovery, I'm doing routine backups, etc...

Short of backing up my system volume and repartitioning, or otherwise starting over, is there a way I can reconfigure my RAID1 array to only expect one disk so that mdadm no longer reports a Degraded state?

View 3 Replies View Related

Fedora :: Rebuild: Cannot Rebuild From Current State

Nov 21, 2010

When trying to rebuild a RAID1 on a FastTrak TX4310 controller after a disk was replaced the below message is given:

[root@SAMBA Samba]# dmraid -R pdc_gigcjhdfa /dev/sdc
Volume "pdc_gigcjhdfa" is not in rebuild state (current: 16)

Rebuild: cannot rebuild from current state!

View 4 Replies View Related

General :: RAID Disk Inactive After Reboot

Aug 12, 2010

I just configured two raid setups but after a reboot they are not mounted and seem to be inactive.

md127 = sde1, sdf1 and sdi1 (raid 5)
md0 = sda1 and sdh1 (raid 0)
root@server /]# cat /proc/mdstat
Personalities :
md127 : inactive sdf1[1](S) sde1[2](S)
78156032 blocks
md0 : inactive sda1[0](S)
488382977 blocks super 1.2
unused devices: <none>

[root@server /]# fdisk -l | grep "Disk /"
Disk /dev/sda: 500.1 GB, 500107862016 bytes
Disk /dev/sdb: 80.0 GB, 80026361856 bytes
Disk /dev/sdc: 122.9 GB, 122942324736 bytes
Disk /dev/sdd: 160.0 GB, 160041885696 bytes
Disk /dev/sde: 40.0 GB, 40020664320 bytes
Disk /dev/sdf: 40.0 GB, 40020664320 bytes
Disk /dev/sdg: 1500.3 GB, 1500301910016 bytes
Disk /dev/sdh: 251.0 GB, 251000193024 bytes
Disk /dev/sdi: 40.0 GB, 40020664320 bytes
Disk /dev/sdj: 500.1 GB, 500107862016 bytes

[root@server /]# cat /etc/mdadm.conf
DEVICE /dev/sdi1 /dev/sdf1 /dev/sde1 /dev/sda1 /dev/sdh1
ARRAY /dev/md127 UUID=5dc0cf7a:8c715104:04894333:532a878b auto=yes
ARRAY /dev/md0 UUID=65c49170:733df717:435e470b:3334ee94 auto=yes

As you can see they now show up as inactive. And for some reason sdi1 and sdh1 are not even listed. What can I do to get them back? To make matters worse I placed some important data on them, and even if I was clever enough to keep an extra copy on another drive, guess which drive that was? So, I need to get them activated as is (at least so I can get the data of them) before I can rebuild them from scratch. I'm running Mandriva 2010.1 and rated tehm using the built in disk partitioner.

View 14 Replies View Related

General :: Setup RAID 5 And One Spare Disk

Aug 4, 2010

I want to build a 6xSATA RAID 5 system with on of the disks as spare disk. I think this give me a chance of 2 of 6 disks failing without losing data. I am right?
Hardware: Intel ICH10R
First I will creat a 3xSATA RAID 5, after I will add the spare disk and after that I will add the others disks. This is what I think I should do.

Step 1:
Create RAID Device
mdadm --create --verbose /dev/md0 --metadata 1.2 --level=5 --raid-devices=3 /dev/sda1 /dev/sdb1 /dev/sdc1
I read that "--metadata 1.2" is the best option. It is true?
Create filesystem on the RAID device

Using this method of calculation:
* chunk size = 128kB (for RAID 5)
* block size = 4kB (recommended for large files, and most of time)
* stride = chunk / block = 128kB / 4k = 32kB
* stripe-width = stride * ( (n disks in raid5) - 1 ) = 32kB * ( (5)- 1 ) = 32kB * 4 = 128kb
mkfs.ext3 -v -m .1 -b 4096 -E stride=32,stripe-width=128 /dev/md0

Step 2:
Add spare-disk
mdadm --add /dev/md0 /dev/sdd1
Is this enough?

Step 3:
Adding disks:
mdadm --add /dev/md0 /dev/sde1
mdadm --grow /dev/md0 --raid-devices=4
fsck.ext3 /dev/md0
resize2fs /dev/md0

View 1 Replies View Related

General :: RAID Disk Failure Interrupt Notification?

Apr 4, 2010

I have installed a Fedora Core 12 Linux system onto a RAID 1 file system. I now need a way of getting an notification if the disk fails. Is there an SNMP MIB that covers Intel RAID? I have done the searching but still the answer alludes me.

View 1 Replies View Related

General :: Debian Software RAID 1- Boot From Both Disk

Mar 15, 2011

I newly installed debian squeeze with software raid. The way I did was, as also given in this thread.

- I have 2 HDD with 500 GB each. For each of them, I created 3 partitions (/boot, / and swap)
- I selected the hard drive and created a new partition table
- I created a new partition that was 1GB. I then specified to use the partition as a Physical Volume for RAID. and used for /boot and enabled bootable.
- Created another partition, which is of 480 GB, and then specified to use the partition as a Physical Volume for RAID. and used for /.
- Created another partion and used for swap

Then RAID configuration:
Through Configure RAID menu -> create MD device ->
(2 for the number of drives, 0 for spare devices)
Next select the partitions you want to be members of /dev/MD0. I selected /dev/sda1 and /dev/sdb1 (for /boot)
Next select the partitions you want to be members of /dev/MD1. I selected /dev/sda6 and /dev/sdb6 (for /)
And no RAID for swap partitions

'Finish Partitioning and write changes to disk' --> Finish the rest of the install like normal. Everything is ok now, except I am not sure how to test my raid config. When I pull the power of the HDD, it only boots from one disk. I read in some forum that I may have to install GRUB manually on the other. In Debian Squeeze, there is no grub command. Not sure how to make my software raid bootable from both disk. I configured /boot partitions of both disks to be boot=yes. Not sure whether that is ok.

View 2 Replies View Related

General :: Where Is The Missing Disk Space On Software Raid

Nov 3, 2010

Purchased (4) 2TiB Drives (actual disk space) and created a RAID5 array expecting to have 6TB of useable disk space, however actual useable space is 5.46TiB.

So, the question is where did the disk space go?

First off, I can say for certainty the disks actual useable is verified at 2TB each have mounted and formated on a non-linux system (OSX).

Disks - 2TB Per disk, Tested HFS, Actual 2TB Useable
root@server:/server# fdisk -l 2>/dev/null | egrep "sd[hijk]" | grep Disk
Disk /dev/sdh: 2000.4 GB, 2000398934016 bytes
Disk /dev/sdj: 2000.4 GB, 2000398934016 bytes
Disk /dev/sdk: 2000.4 GB, 2000398934016 bytes


View 2 Replies View Related

General :: MDADM Error: Creating A RAID 1 RAM Disk?

Apr 7, 2011

I am trying to create a Raid 1 ram disk. Below are the commands I used:

[root@abidbodal dev]# mke2fs -m 0 /dev/ram8
[root@abidbodal dev]# mount /dev/ram8 /mnt/rd8
[root@abidbodal dev]# mke2fs -m 0 /dev/ram9


View 3 Replies View Related

Hardware :: Fedora 11 RAID 1 - Disk Failure - Boot From The Single Working Disk?

Oct 16, 2009

my Fedora 11 system is not starting anylonger. It stops with the message:


VFS: Can't find ext4 filesystem on dev dm-0

The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).

View 2 Replies View Related

Ubuntu Servers :: HW RAID Disk Shows Up In Fstab But Not In /dev/disk/by-uuid?

Jun 28, 2010

I have an SiI hardware SATA RAID card, with two 500GB disks in mirrored RAID configuration. When I first plugged them in and set it up, things seemed to work ok, but on boot the raid controller told me that the RAID needed rebuilding, and it would happen automatically after POST. So I didn't worry about it, and the drive mounted fine, and it's been that way for years. I just went in and manually on-line rebuilt the RAID in the controller's BIOS, and now when I boot into Ubuntu, both disks show up in fdisk, but neither show up in /dev/disk/by-uuid. Am I missing something?

View 9 Replies View Related

General :: Mount A Single RAID 1 Disk / Partition As Ext3?

Jul 7, 2011

I need to copy data from a single HD, which used to be part of a Linux RAID 1. I've googled around, but can't find any clue how to mount partitions from this single HD.

Background: The HD comes from a linux based NAS box Synology DS207+. The NAS uses ext3 as filesystem. Both NAS disks are fine, but the other NAS hardware is dead and not worth repairing or replacing.

View 1 Replies View Related

Ubuntu Installation :: Migrate Working Single Disk System To Existing RAID Array Using Disk UUIDs

Aug 1, 2010

I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.

I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).

These are the commands I used:


p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*


I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...

So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.

It's still trying to use /dev/disk/by-uuid/412d... to boot.

What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.

View 2 Replies View Related

General :: Scsi RAID Jbod And Arrays - Disk Utilization And The Corresponding Low Data Transfer

Jul 6, 2010

So I have a system that is about 6 years old running Redhat 7.2 that is supporting a very old app that cannot be replaced at the moment. The jbod has 7 Raid1 arrays in it, 6 of which are for database storage and another for the OS storage. We've recently run into some bad slowdowns and drive failures causing nearly a week in downtime. Apparently none of the people involved, including the so-called hardware experts could really shed any light on the matter. Out of curiosity I ran iostat one day for a while and saw numbers similar to below:


Some of these kinda weird me out, especially the disk utilization and the corresponding low data transfer. I'm not a disk IO expert so if there are any gurus out there willing to help explain what it is I'm seeing here. As a side note, the system is back up and running it just runs sluggish and neither the database folks nor the hardware guys can make heads or tails of it. Ive sent them the same graphs from iostat but so far no response.

View 1 Replies View Related

Ubuntu :: Automatic Disk Check Missing?

Apr 19, 2011

Ubuntu has got this build-in check for errors which starts every 30 startups (if I remember well ) but my one gone missing... Strange. How can I turn it back on ?ound in the forum some information about Bonager, but is this original automatic disk check software shipped with Ubuntu or another piece of software ?

View 5 Replies View Related

Copyrights 2005-15, All rights reserved