CentOS 5 :: Software RAID 1 HDD Failure
Nov 6, 2009
1. One of my hdds failed (sda) in software raid 1. I rma'd the hdd to western digital and got another one. Now do I have to format it before putting it in my centos server? If yes, how do I format?
2. Also since sda drive failed, I gotta mark sda as failed in raid. Then remove the sda hdd, and pop in the new hdd for sda? Or do I switch sdb to sda and put the new hdd in sdb's place?
3. After that add it to raid correct, then once raid rebuilds I have to do grub? Can grub be done via ssh only? or do I need to be at the datacenter or get kvm?
4. Last question, I've got a supermicro hotswap hdd case, so do I need to shutdown server while I replace the hdd's? I just want to be sure I do this correctly.
The following is the guide that I will be using, please look at it and let me know if that is the correct procedure: [URL]. Another thing, when the hdd (sda) failed I put it back into raid, but the hdd has bad sectors that is why m replacing it.
View 8 Replies
ADVERTISEMENT
Jun 9, 2011
Following scenario: My server in some data center on a different continent with two disks and software raid 1.
One day I see that a disk failed (for example with /proc/mdstat). Of course I should replace the failed disk asap. Now that I think about it, I am not sure how. What should my email to the data center support guy mention to make sure that guy doesn't replace the wrong disk?
With hardware RAID it is very easy, because the controller usually has some kind of red LED indicator. But what about software raid?
View 8 Replies
View Related
Oct 12, 2010
Recently while using a Highpoint 2310 (raid 5) I lost the mother board and CPU. I had to reinstall Centos and found it needed to initialize the array to function. Total loss of date. Question: If I use a true hardware card (3ware 9650se) and experience a serious hardware loss or the C drive can the card be installed with the drives on a new motherboard and function without data loss even if the OS must be reinstalled.
View 4 Replies
View Related
Apr 28, 2009
I've just setup a new centos 5.3 server with a 3-disc raid 5 software RAID array. I've setup other software raid 5 arrays on this same hardware, whilst testing, and had no trouble... I only just installed 3 new drives and performed a new install from scratch.
Hardware is: 4800XP X2 64-bit, 2GB RAM, Albatron KI-690G mainboard with Marvel SATA controller (I think) - 4 ports.
SATA port 0 is system drive (OCZ vertex 32GB SSD)
SATA port 1 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 2 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 3 is Western Digital "green" 1TB 8MB cache SATA 2
Ports 1-3 in software raid MD0 (raid 5)
All of this was configured in the GUI setup, with LVM on top of software RAID 5 mounted on /var. Partition size is ~1.9TB
Trouble is I get all sorts of "hardware failure" messages at bootup and the MD driver reports it was only able to bring up 2 out of 3 drives in the RAID set... however the RAID set formatted fine during the setup?
Here is the relevent dmesg output...
View 2 Replies
View Related
Jun 5, 2011
I have 4 WD10EARS drives running in a RAID 5 array using MDADM.Yesterday my OS Drive failed. I have replaced this and installed a fresh copy of Ubuntu 11.04 on it.I then installed MDADM, and rebooted the machine, hoping that it would automatically rebuild the array.It hasnt, when i look at the array using Disk Utility, it says that the array is not running. If i try to start the array it says : Error assembling array: mdadm exited with exit code 1: mdadm: failed to RUN_ARRAY /dev/md0: Input/output errormdadm: Notenough devices to start the array.I have tried MDADM --assemble --scan and it gives this output:mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.I know that there are 4 drives present as they are all showing, but it is only using 2 of them.I also ran MDADM -- detail /dev.md0 which gave:
root@warren-P5K-E:~# mdadm --detail /dev/md0
/dev/md0:
Version : 0.90
[code]....
View 2 Replies
View Related
May 22, 2009
I *had* a server with 6 SATA2 drives with CentOS 5.3 on it (I've upgraded over time from 5.1). I had set up (software) RAID1 on /boot for sda1 and sdb1 with sdc1, sdd1, sde1, and sdf1 as hot backups. I created LVM (over RAID5) for /, /var, and /home. I had a drive fail last year (sda).After a fashion, I was able to get it working again with sda removed. Since I had two hot spares on my RAID5/LVM deal, I never replaced sda. Of course, on reboot, what was sdb became sda, sdc became sdb, etc.So, recently, the new sdc died. The hot spare took over, and I was humming along. A week later (before I had a chance to replace the spares, another died (sdb).Now, I have 3 good drives, my array has degraded, but it's been running (until I just shut it down to tr y.
I now only have one replacement drive (it will take a week or two to get the others).I went to linux rescue from the CentOS 5.2 DVD and changed sda1 to a Linux (as opposed to Linux RAID) partition. I need to change my fstab to look for /dev/sda1 as boot, but I can't even mount sda1 as /boot. What do I need to do next? If I try to reboot without the disk, I get insmod: error inserting '/lib/raid456.ko': -1 File existsAlso, my md1 and md2 fail because there are not enough discs (it says 2/4 failed). I *believe* that this is because sda, sdb, sdc, sdd, and sde WERE the drives on the raid before, and I removed sdb and sdc, but now, I do not have sde (because I only have 4 drives) and sdd is the new drive. Do I need to label these drives and try again? Suggestions? (I suspect I should have done this BEFORE failure).Do I need to rebuild the RAIDs somehow? What about LVM?
View 6 Replies
View Related
Aug 3, 2010
I am trying to connect a RAID Box to the server via LSI 8880EM2 RAID controller.The raid box is FUjitsu Externus DX60 with raid 1 configured.The server is Fujitsu Primergy SX300 S5 with LSI 8880EM2 RAID card.The external raid box is being recognised by the raid controllers bios.
The server runs CentOS 5.5 64bit. I have installed the megaraid_sas driver from LSI website and MegaCLI utility but CentOS still fails to see the box.MegaCLI utility, when launched from CentOS, recognises the raid box, CentOS does not(no mapping being created in /dev).
I have also tried to create a logical RAID0 HDD on one physical HDD(as seen by MegaCLI) with MegaCLI utility in CentOS.The result was success and the new logical drive could have been used, but when restarting the server, the controllers' bios fails with an error (not surprised(one logical RAID0 on one physical HDD)) and configuration is being erased.
Has anyone tried connecting 8880EM2 controller to a raid box with raid configured on the box, running it all under CentOS and what were the results.
View 3 Replies
View Related
Feb 2, 2010
Recently, one the SMART utility said that one of the drives had failed and another drive was about to fail. I downed the box and hooked them up to my windows machine to run sea tools on them (They are all seagate drives). Sea Tools said that the drives were fine, while ubuntu said they were failing/dead. Yesterday I decided to try to fix one of the drives in the raid. I turned the server off, took the failed drive out, and restarted. Of course the raid didn't work because only 2 of the 3 drives were there, however it had been working w/ only 2 of the 3 drives for a couple months now (I'm a lazy college student). I turned it back off and back on with the drive there just to see if I could get the raid up again, but I havn't been able to get it to go. So far I've tried:
Code:
mdadm --assemble /dev/md0 /dev/sd[b,c,d]
mdadm: no recogniseable superblock on /dev/sdb
mdadm: /dev/sdb has no superblock - assembly aborted
[code]....
I'm looking for a way to trick the raid into working with just 2 drives until I can warranty the seagate and buy an external 1.5 TB drive to use as another backup. how to remove the bad drive from the array and replace it with a fresh drive, without data loss.
View 3 Replies
View Related
Jun 18, 2010
I have a fileserver which is running Ubuntu Server 6.10. I had a RAID5 array consisting of the following disks:
Code:
/dev/sda1
/dev/sdb1
/dev/sdd1
/dev/md0 -
the raid drive for the above three disks. The sda1 disk has failed and the array is running on 2 of 3 disks
/dev/sdc (OS disk)
/dev/sde (new 2tb disk - unused)
/dev/sdf (new 2tb disk - unused)
My plan was to rebuild the array using the two new disks as RAID1. Would the best way to do this be to create a new RAID1 disk on /dev/md1 then copy all data over from /dev/md0? Also - this may sound stupid but since all 3 drives in md0 are identical i'm not sure physically which disk is bad. I tried disconnecting each disk one-by-one then rebooting but the system doesn't appear to want to boot without the bad drive connected. I've already failed the disk in the array with mdadm but i'm unsure of how to remove it properly.
View 3 Replies
View Related
Sep 6, 2010
Based on the reading I've done over the past 48 hours I think I'm in serious trouble here with my RAID 5 array. I got another 1 TB drive and added to my other 3 to increase my space to 3 TB...no problem.
While the array was resyncing...it got to about 40%, I had a power failure. So I'm pretty sure it failed while it was growing the array...not the partition. Next time I booted mdadm didn't even detect the array. I fiddled around trying to get mdadm to recognize my array, but no luck.
I finally got desperate enough to just create the array again...I knew the settings of my and had seen some people have success with this method. When creating it, it asked me if I was sure because the disks appeared to belong to an array already, but I said yes. The problem is when I created it, it created a clean array and this is what I'm left with.
Code:
/dev/md0:
Version : 00.90
Creation Time : Sun Sep 5 20:01:08 2010
Raid Level : raid5
Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
[Code]....
I tried looking for backup superblock locations using e2fsck and every other tool I could find, but nothing worked. I tried testdisk which says it found my partition on /dev/md0, so I let it create the partition. Now I have a /dev/md0p1, which won't let me mount it either. What's interesting is gparted reports /dev/md0p1 as the old partition size (1.82 TB)...the data has to still be there, right?
View 3 Replies
View Related
Sep 13, 2010
on my pc, that was running WinXP, I thought of installing Ubuntu. (I did install linux already a few times in the past years and use it on another couple of pcs) But something went wrong. This machine has 2 x 200MB maxtor drives, in raid 0 configuration, supported by the motherboard Nvidia chipset, and working well in Windows. When ran the live Ubuntu 10.04 cd, gparted was not able to access the drives in raid configuration, until I installed the mdadm and kpartx packages then the existing data became visible. So after that initial moment I thought all was ok and proceeded to install Lucid on the machine, dual booting with Windows. I did partition manually so that in my 400GB raid drive there is an 80GB NTFS partition with WinXP, a 90GB extended partition for Linux Ext4 and Swap and then a last NTFS 200GB partition for data. All went well, but now on restarting the computer nothing happens, nothing loads, Grub is not showing, and it looks like I cannot launch Linux nor Windows. All the data from WinXP and the Ubuntu installation seems to be on the disks but the pc is just not booting. I suppose the problem is with the raid configuration that is not handled properly during the installation, but is there anything that I can do now, apart from reinstalling Windows Xp or installing Ubuntu in a non raid configuration?
View 9 Replies
View Related
Dec 19, 2010
I went to setup my linux box and found that the OS drive had finally died. It was an extremely old WD raptor drive in a hot box full of drives so it was really only a matter of time before it just quit on me. Normally this wouldn't be such a big deal however I had just recently constructed an md RAID5 array of 3 1TB disks to act as an NFS mount for basically all of my important files. Maybe 2-3 weeks before the failure I had finished moving all of my most important stuff onto that array. Now I know that the array is intact. All the required data is sitting on those disks. Since only the OS level disk failed on me I should be able to get a new disk in there, reinstall ubuntu and then rebuild that array. how exactly do I go about doing that with mdadm? Do I create the array from the /dev character devices like when I initially built the array?
View 2 Replies
View Related
Apr 4, 2010
I have installed a Fedora Core 12 Linux system onto a RAID 1 file system. I now need a way of getting an notification if the disk fails. Is there an SNMP MIB that covers Intel RAID? I have done the searching but still the answer alludes me.
View 1 Replies
View Related
Jun 7, 2011
I have 4 WD10EARS drives running in a RAID 5 array using MDADM.Yesterday my OS Drive failed. I have replaced this and installed a fresh copy of Ubuntu 11.04 on it. then installed MDADM, and rebooted the machine, hoping that it would automatically rebuild the array.It hasnt, when i look at the array using Disk Utility, it says that the array is not running. If i try to start the array it says :Error assembling array: mdadm exited with exit code 1: mdadm: failed to RUN_ARRAY /dev/md0: Input/output error
mdadm: Not enough devices to start the array.I have tried MDADM --assemble --scan and it gives this output:mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.I know that there are 4 drives present as they are all showing, but it is only using 2 of them.I also ran MDADM -- detail /dev.md0 which gave:
root@warren-P5K-E:~# mdadm --detail /dev/md0
/dev/md0:
Version : 0.90
[code]...
View 11 Replies
View Related
May 1, 2011
I just setup sendmail on my server to send emails and it works, now I would like to be able to get an email from mdadm if sometjhing was going wrong. I imagine most raid users have this feature setup.
Right now, I have 7 raid arrays and mdadm starts at boot time. Until now, I used Mr. Goblin's script (http://connie.slackware.com/~mrgoblin/files/rc.mdadm) (thanks Mr Goblin!) to monitor my arrays.
The script is started at boot time from rc.local. I created a small script in /usr/bin that send the following command to rc.mdadm giving me the status of the arrays:
Code:
/etc/rc.d/rc.mdadm status
and it works fine, but this requires me probing the arrays manually by calling the script from the command line. I would like to automate probing every 10 minutes or whatever and if a fault has been detected, I get an email.
[Code]...
View 14 Replies
View Related
Aug 27, 2010
UPDATE: decided to reinstall and run the partitioner to get rid of the raid. Not worth dealing with this since seems to be lower level as /dev/mapper was not listing any devices. Error 15 at grub points to legacy grub. So avoiding the problem by getting rid of raid for now. So ignore this post. Found a nice grub2 explanation on the wiki but didn't help this situation since probably isn't a grub problem. Probably is a installer failure to map devices properly when it only used what was already available and didn't create them during the install. I don't know, just guessing. Had OpenSuSE 10.3 64bit installed with software raid mirrored swap, boot, root. Used the alternate 64bit Ubuntu iso for installation. Since partitioning was already correctly setup and the raid devices /dev/md0,1,2 were recognized by the installer, I chose to format the partitions with ext3 and accept the configuration:
/dev/md0 = swap = /dev/sda1, /dev/sdb1 = 2Gb
/dev/md1 = boot = /dev/sda2, /dev/sdb2 = 256Mb
/dev/md2 = root = /dev/sda3, /dev/sda3 = 20Gb
Installation process failed at the point of installing grub. It had attempted to install the bootloader on /dev/sda2 and /dev/sdb2. I moved on since it would not let me fiddle with the settings and I got the machine rebooted with the rescue option on the iso used for installing. Now, I can see the root partition is populated with files as expected. dpkg will list that linux-image-generic, headers, and linux-generic are installed with other supporting kernel packages. grub-pc is installed as well. However, the /boot partition or /dev/md1 was empty initially after the reboot. What is the procedure to get grub to install the bootloader on /dev/sda2 and /dev/sdb2, which represent /dev/md1 or /boot?
Running apt-get update and apt-get upgrade installed a newer kernel and this populated the /boot partition. Running update-grub results in a "/usr/sbin/grub-probe: error: no mapping exists for 'md2'". grub-install /dev/md2 or grub-install /dev/sda2 gives the same error as well. Both commands indicate that "Autodetection of a filesystem module failed, Please specify the module with the option '--modules' explicitly". What is the right modules that need to be loaded for a raid partition in initrd? Should I be telling grub to use the a raid module?
View 1 Replies
View Related
Jan 3, 2009
I had a RAID1 'device' build on two physical partitions on two drives. One of the disk controllers died and software RAID did the job - now I am working on the degraded array.
Now I want to put the old disk (sdb) back, and I am not sure what will happen. Both disks have 'raid auto' partitions. And sdb file structure from before of the failure. The raid code will find inconsistency between both partitions. What will it decide? Will it start coping from the currently running system (sda) all the data to the old one (sdb) at the boot time, as I wish?
I don't want to it to write from the old one to the new one, as some months passed and lots of changes happened to the data.
View 1 Replies
View Related
Dec 10, 2009
I am going to be using CentOs 5.4 for a home storage server. It will be RAID6 on 6 x 1TB drives. I plan on using an external enclosure which is connected via two SFF-8088 cables (4 drives a piece). I am looking to try and find a non-RAID HBA which would support this external enclosure and allow to use standard linux software raid.
If this is not an option, I'd consider using a hardware based raid card, but they are very expensive. The Adaptec 5085 is one option but is almost $800. If that is what I need for this thing to be solid then that is fine, I will spend the money but I am thinking that software raid may be the way to go.
View 3 Replies
View Related
Dec 7, 2010
I'm working on a server and noticed that the to RAID5 setup is showing 4 Raid devices but only 3 Total devices. It's on a fully updated CentOS 5 system that only has three SATA drives, as it can not hold anymore. I've done some researching but am unable to remove the fourth device, which is listed as removed. The full output of `mdadm -D /dev/md2` can be see below. I've never run into this situation before.Anyone have any pointers on how I can reduced the Raid Devices from 4 to 3? I have tried
mdadm /dev/md2 -r failed
mdadm /dev/md2 -r detached
but neither work and since there is no block device listed I'm not quite sure how to get things back in sync so it's only seeing the three drives.
/dev/md2:
Version : 0.90
Creation Time : Tue May 25 11:07:04 2010
Raid Level : raid5
[code]....
View 8 Replies
View Related
Oct 16, 2009
my Fedora 11 system is not starting anylonger. It stops with the message:
Code:
VFS: Can't find ext4 filesystem on dev dm-0
The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).
View 2 Replies
View Related
Jan 19, 2011
I have (had) Debian Testing running on a 250GB IDE hard drive, partitioned normally.
I also have 4x 1TB drives in a raid 5 using mdadm, and 2x 500GB drives in a raid 1 also with mdadm.
I put the two arrays in lvm using:
I then used "lvcreate" to make storage/backup 300GB, and the rest went to storage/media (approx. 2TB usable). I put an xfs filesystem on both and mounted them.
All was working fine until the system drive shorted out and died on me this morning. As far as I can tell, all my other drives and everything else is fine. I do a daily rsnapshot of the filesystem, which of course is residing on storage/backup (stupid, I know). So I have full backups of everything, but I'll have to put a new hard drive in and reinstall Debian before I can restore everything.
I've reinstalled before and simply reassembled mdadm arrays and remounted them before with no problems, but this is the first time I've used lvm, so I'm not sure what I have to do to restore everything. Is it as simple as reinstalling the system then doing a:
View 4 Replies
View Related
Feb 15, 2009
The installer can't see my raid controller (I assume) as I'm getting the following error:"Error opening /dev/mapper/isw_jbhgjgjj_Vol0: No such device or address"It just sees them as 4 individual drives: sda, sdb, sdc and sdd.Please note that I have set up the RAID 5 in the controller bios interface and the image name is Vol0, which it seems that it tries to load but for some particular reason it can't.I have also tried different bios settings and nothing worked.
View 3 Replies
View Related
Jan 13, 2010
My problem is that I'm trying to install CentOS 5.4 x86_64 DVD ISO on Supermicro X7SBI server with installed Adaptec RAID 3405 controller.
I created RAID 5 array and is working fine (adaptec status says Optimal) but I can't install CentOS to that array (1.5TB size).
Whenever I try to install with: linux dd
I'm asked for a driver, which I have downloaded from Adaptec site and extracted contents to USB drive (in installation found as /sba1) which has now a lot of IMG and some ISO files on it.
I try to load (I simplified names) RHEL5.img, CENTOS.img... with x64 names (one exact name: aacraid driverdisk-CentOS-x86_64.img) and I always get the error message: "No devices of the appropriate type were found on this driver disk"
This is going on for a week now and I can't find the right driver or something I'm doing wrong to get install done.
View 16 Replies
View Related
May 7, 2009
I need your help urgently. I will shortly install Centos 5 on a HP DL380R05 E5420 Server with two HP 146GB 10k 2.5 SAS HP SP hard disks running hardware RAID-1. I am a newbie on CentOS, please advise me:
1. Do I need a Centos RAID device driver for hardware RAID-1 on the HP DL380R05 E5420 Server?
2. Should I use Centos i386 or Centos x86_64 in this case? Which one: CentOS 5.2 or CentOS 5.3?
3. Does Centos support Embedded NC373i Multifunction Gigabit network cards?
The following is my server's hardware configuration:
HP DL380R05 E5420 2GB Base AP Server(1) Quad-Core Intel Xeon Processor E5420 Two Embedded NC373i Multifunction Gigabit / 2 GB (2 x 1 GB) PC2-5300 Fully Buffered DIMMs (DDR2-667)/HP Smart Array P400/256MB Controller /Rack (2U)/Hot Plug Fully Redundant Fans Standard Two HP 146GB 10k 2.5 SAS HP SP HDD HP 1000W IEC C14 Cord RPS Kit
View 19 Replies
View Related
Dec 31, 2009
I've a Centos 5.1 server on S5000vsa Intel board with 6 bay HD in raid 1 mode (2 HD), and working fine until I've decide to make a yum update. Today the Centos is 5.4 but seems necesssary to upgrade the megasr driver for the raid function (i.e. , is not possible to contact the raid firmware trought Web Console 2). I've download the last megasr driver from Intel and I follow the instruction to upgrade the driver with s.o. installed, but when I start the server and choose via grub the new img boot file, the display show :
megasr.ko invalid module format -1 and stop. The linux version is the same in the machine (uname -a) and from the driver (2.6.18-164.9.1.el5). Another question : in Intel instruction spoken about megasr.o , but on img file I found a megasr.ko . I think instructions referre to old linux 2.4
View 3 Replies
View Related
Oct 16, 2010
I'm setting up a raid 5 on several hard disks with a layer of lvm on top for good measure.I know the recent kernels support growing software raid, but since centos runs 2.6.18, I wanted to make sure it'll work. Does the centos kernel support growing raid devices?
View 1 Replies
View Related
Feb 1, 2011
Could any RAID gurus kindly assist me on the following RAID-5 issue?I have an mdadm-created RAID5 array consisting of 4 discs. One of the discs was dropping out, so I decided to replace it. Somehow, this went terribly wrong and I succeeded in marking two of the drives as faulty, and the re-adding them as spare.
Now the array is (logically) no longer able to start:
mdadm: Not enough devices to start the array.Degraded and can't create RAID ,auto stop RAID [md1]
I was able to examine the disks though:
Code:
root@127.0.0.1:/etc# mdadm --examine /dev/sdb2
/dev/sdb2:
Magic : a92b4efc
Version : 00.90.00
code....
Code:
mdadm --create --assume-clean --level=5 --raid-devices=4 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2
As I don't want to ruin the maybe small chance I have left to rescue my data, I would like to hear the input of this wise community.
View 4 Replies
View Related
May 10, 2011
I have DELL R310 server with PERC 200H controller and RAID 5 configured already.This is my first time I want to install CentOS and seems like a new challenge for me I've read some articles on Wiki and posts on this forum and didn't find any information how to install CentOS on RAID 5.I only noticed that if there is fake or software RAID1 it should be switched off and made by OS configuration.I think PERC 200H is separate controller so RAID 5 is hardware RAID, is there anybody who can confirm ?My problem is that when I try to install CentOS it doesn't see any hard disks ... I looked for drivers on DELL web site and there is driver for Red Hat 6. I tried to install it using linux dd command but it doesn't recognize. Is there anybody else who met such problem ?
View 3 Replies
View Related
Apr 1, 2009
Just tried updating 5.2 to 5.3; using both pup and yum, and both get the same error:
Error: Missing Dependency: libgpod.so.0 is needed by package rhythmbox
If I remove rhythmbox (I don't use it on this machine) the upgrade works, but I thought I'd post in case someone wanted to take a look at it.
View 4 Replies
View Related
Apr 1, 2009
At 4:00 pm on 4/1/2009 I accepted the offer to dowload and install a number of small improvements. This has always worked in the past. This time there were a few problems with the install, but I went ahead and rebooted. It stopped at grub, stage2.
I am now trying the rescue feature of my install CD. I will report on this later.
As for the always possible hardware failure, I have a raid setup, so I was not expecting this.
View 4 Replies
View Related