CentOS 5 :: Disk Failure And Software Raid 1?
Jun 9, 2011
Following scenario: My server in some data center on a different continent with two disks and software raid 1.
One day I see that a disk failed (for example with /proc/mdstat). Of course I should replace the failed disk asap. Now that I think about it, I am not sure how. What should my email to the data center support guy mention to make sure that guy doesn't replace the wrong disk?
With hardware RAID it is very easy, because the controller usually has some kind of red LED indicator. But what about software raid?
View 8 Replies
ADVERTISEMENT
Oct 16, 2009
my Fedora 11 system is not starting anylonger. It stops with the message:
Code:
VFS: Can't find ext4 filesystem on dev dm-0
The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).
View 2 Replies
View Related
Feb 2, 2010
Recently, one the SMART utility said that one of the drives had failed and another drive was about to fail. I downed the box and hooked them up to my windows machine to run sea tools on them (They are all seagate drives). Sea Tools said that the drives were fine, while ubuntu said they were failing/dead. Yesterday I decided to try to fix one of the drives in the raid. I turned the server off, took the failed drive out, and restarted. Of course the raid didn't work because only 2 of the 3 drives were there, however it had been working w/ only 2 of the 3 drives for a couple months now (I'm a lazy college student). I turned it back off and back on with the drive there just to see if I could get the raid up again, but I havn't been able to get it to go. So far I've tried:
Code:
mdadm --assemble /dev/md0 /dev/sd[b,c,d]
mdadm: no recogniseable superblock on /dev/sdb
mdadm: /dev/sdb has no superblock - assembly aborted
[code]....
I'm looking for a way to trick the raid into working with just 2 drives until I can warranty the seagate and buy an external 1.5 TB drive to use as another backup. how to remove the bad drive from the array and replace it with a fresh drive, without data loss.
View 3 Replies
View Related
Jun 18, 2010
I have a fileserver which is running Ubuntu Server 6.10. I had a RAID5 array consisting of the following disks:
Code:
/dev/sda1
/dev/sdb1
/dev/sdd1
/dev/md0 -
the raid drive for the above three disks. The sda1 disk has failed and the array is running on 2 of 3 disks
/dev/sdc (OS disk)
/dev/sde (new 2tb disk - unused)
/dev/sdf (new 2tb disk - unused)
My plan was to rebuild the array using the two new disks as RAID1. Would the best way to do this be to create a new RAID1 disk on /dev/md1 then copy all data over from /dev/md0? Also - this may sound stupid but since all 3 drives in md0 are identical i'm not sure physically which disk is bad. I tried disconnecting each disk one-by-one then rebooting but the system doesn't appear to want to boot without the bad drive connected. I've already failed the disk in the array with mdadm but i'm unsure of how to remove it properly.
View 3 Replies
View Related
Dec 19, 2010
I went to setup my linux box and found that the OS drive had finally died. It was an extremely old WD raptor drive in a hot box full of drives so it was really only a matter of time before it just quit on me. Normally this wouldn't be such a big deal however I had just recently constructed an md RAID5 array of 3 1TB disks to act as an NFS mount for basically all of my important files. Maybe 2-3 weeks before the failure I had finished moving all of my most important stuff onto that array. Now I know that the array is intact. All the required data is sitting on those disks. Since only the OS level disk failed on me I should be able to get a new disk in there, reinstall ubuntu and then rebuild that array. how exactly do I go about doing that with mdadm? Do I create the array from the /dev character devices like when I initially built the array?
View 2 Replies
View Related
Apr 4, 2010
I have installed a Fedora Core 12 Linux system onto a RAID 1 file system. I now need a way of getting an notification if the disk fails. Is there an SNMP MIB that covers Intel RAID? I have done the searching but still the answer alludes me.
View 1 Replies
View Related
Nov 6, 2009
1. One of my hdds failed (sda) in software raid 1. I rma'd the hdd to western digital and got another one. Now do I have to format it before putting it in my centos server? If yes, how do I format?
2. Also since sda drive failed, I gotta mark sda as failed in raid. Then remove the sda hdd, and pop in the new hdd for sda? Or do I switch sdb to sda and put the new hdd in sdb's place?
3. After that add it to raid correct, then once raid rebuilds I have to do grub? Can grub be done via ssh only? or do I need to be at the datacenter or get kvm?
4. Last question, I've got a supermicro hotswap hdd case, so do I need to shutdown server while I replace the hdd's? I just want to be sure I do this correctly.
The following is the guide that I will be using, please look at it and let me know if that is the correct procedure: [URL]. Another thing, when the hdd (sda) failed I put it back into raid, but the hdd has bad sectors that is why m replacing it.
View 8 Replies
View Related
Oct 12, 2010
Recently while using a Highpoint 2310 (raid 5) I lost the mother board and CPU. I had to reinstall Centos and found it needed to initialize the array to function. Total loss of date. Question: If I use a true hardware card (3ware 9650se) and experience a serious hardware loss or the C drive can the card be installed with the drives on a new motherboard and function without data loss even if the OS must be reinstalled.
View 4 Replies
View Related
Mar 8, 2010
I've tried to type this message twice in detail, but the computer won't submit it. What are the proper steps to take when a SCSI drive fails in a RAID five array running CentOS 5.4?
Yes, I read the manual. On a Dell 6650 system the drive can't be rebuilt because it is not recognized as the correct size. Linux rescue doesn't work... won't find the linux partitions that are still there according to fdisk -l
View 7 Replies
View Related
Apr 28, 2009
I've just setup a new centos 5.3 server with a 3-disc raid 5 software RAID array. I've setup other software raid 5 arrays on this same hardware, whilst testing, and had no trouble... I only just installed 3 new drives and performed a new install from scratch.
Hardware is: 4800XP X2 64-bit, 2GB RAM, Albatron KI-690G mainboard with Marvel SATA controller (I think) - 4 ports.
SATA port 0 is system drive (OCZ vertex 32GB SSD)
SATA port 1 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 2 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 3 is Western Digital "green" 1TB 8MB cache SATA 2
Ports 1-3 in software raid MD0 (raid 5)
All of this was configured in the GUI setup, with LVM on top of software RAID 5 mounted on /var. Partition size is ~1.9TB
Trouble is I get all sorts of "hardware failure" messages at bootup and the MD driver reports it was only able to bring up 2 out of 3 drives in the RAID set... however the RAID set formatted fine during the setup?
Here is the relevent dmesg output...
View 2 Replies
View Related
Mar 13, 2011
I have a dedicated server that has CentOS 5.5 installed.. I can access that server via SSH as root. Now the issue is.. httpdocs folder is situated in /var where all website data is stored. I have more than 50GB of website that needs to be transfered to this partition but this partition is of 4.0GB..
# cat /etc/fstab
/dev/md1 / ext3 defaults 1 1
/dev/sda2 none swap sw
/dev/sdb2 none swap sw
[code]...
View 2 Replies
View Related
Nov 17, 2009
Our server is a CybertronPC I2XV9080 Imperium Tower. It is equipped with a supermicro X7DVL-I Motherboard and Quad 750 GB SATA2 RAID edition hard drives in a raid 5 array. We tried to install Centos on the Raid5 array with Device-Mapper as the LVM. In the BIOS SATA Raid was enabled and the ICH RAID code base option was set to [Intel].
Intel Matrix Storage Manager Option ROM V5.6.4.1002 ESB2
RAID
ID Name Level Strip Size Status Bootable
0 Raid5 Raid 5 64KB 80GB Normal Yes
1 Raid_5 Raid 5 64kB 2000GB Normal Yes[code].....
Can I have multiple level raids across the same array or would that lead to problems as above? Is the root cause of my problem the fact that intel raid5 is not supported for Linux as based on the following link http:[url]....
View 3 Replies
View Related
Jun 28, 2010
I have an SiI hardware SATA RAID card, with two 500GB disks in mirrored RAID configuration. When I first plugged them in and set it up, things seemed to work ok, but on boot the raid controller told me that the RAID needed rebuilding, and it would happen automatically after POST. So I didn't worry about it, and the drive mounted fine, and it's been that way for years. I just went in and manually on-line rebuilt the RAID in the controller's BIOS, and now when I boot into Ubuntu, both disks show up in fdisk, but neither show up in /dev/disk/by-uuid. Am I missing something?
View 9 Replies
View Related
Aug 1, 2010
I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.
I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).
These are the commands I used:
Quote:
p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*
[Code]....
I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...
So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.
It's still trying to use /dev/disk/by-uuid/412d... to boot.
What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.
View 2 Replies
View Related
Jun 5, 2011
I have 4 WD10EARS drives running in a RAID 5 array using MDADM.Yesterday my OS Drive failed. I have replaced this and installed a fresh copy of Ubuntu 11.04 on it.I then installed MDADM, and rebooted the machine, hoping that it would automatically rebuild the array.It hasnt, when i look at the array using Disk Utility, it says that the array is not running. If i try to start the array it says : Error assembling array: mdadm exited with exit code 1: mdadm: failed to RUN_ARRAY /dev/md0: Input/output errormdadm: Notenough devices to start the array.I have tried MDADM --assemble --scan and it gives this output:mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.I know that there are 4 drives present as they are all showing, but it is only using 2 of them.I also ran MDADM -- detail /dev.md0 which gave:
root@warren-P5K-E:~# mdadm --detail /dev/md0
/dev/md0:
Version : 0.90
[code]....
View 2 Replies
View Related
May 22, 2009
I *had* a server with 6 SATA2 drives with CentOS 5.3 on it (I've upgraded over time from 5.1). I had set up (software) RAID1 on /boot for sda1 and sdb1 with sdc1, sdd1, sde1, and sdf1 as hot backups. I created LVM (over RAID5) for /, /var, and /home. I had a drive fail last year (sda).After a fashion, I was able to get it working again with sda removed. Since I had two hot spares on my RAID5/LVM deal, I never replaced sda. Of course, on reboot, what was sdb became sda, sdc became sdb, etc.So, recently, the new sdc died. The hot spare took over, and I was humming along. A week later (before I had a chance to replace the spares, another died (sdb).Now, I have 3 good drives, my array has degraded, but it's been running (until I just shut it down to tr y.
I now only have one replacement drive (it will take a week or two to get the others).I went to linux rescue from the CentOS 5.2 DVD and changed sda1 to a Linux (as opposed to Linux RAID) partition. I need to change my fstab to look for /dev/sda1 as boot, but I can't even mount sda1 as /boot. What do I need to do next? If I try to reboot without the disk, I get insmod: error inserting '/lib/raid456.ko': -1 File existsAlso, my md1 and md2 fail because there are not enough discs (it says 2/4 failed). I *believe* that this is because sda, sdb, sdc, sdd, and sde WERE the drives on the raid before, and I removed sdb and sdc, but now, I do not have sde (because I only have 4 drives) and sdd is the new drive. Do I need to label these drives and try again? Suggestions? (I suspect I should have done this BEFORE failure).Do I need to rebuild the RAIDs somehow? What about LVM?
View 6 Replies
View Related
Aug 3, 2010
I am trying to connect a RAID Box to the server via LSI 8880EM2 RAID controller.The raid box is FUjitsu Externus DX60 with raid 1 configured.The server is Fujitsu Primergy SX300 S5 with LSI 8880EM2 RAID card.The external raid box is being recognised by the raid controllers bios.
The server runs CentOS 5.5 64bit. I have installed the megaraid_sas driver from LSI website and MegaCLI utility but CentOS still fails to see the box.MegaCLI utility, when launched from CentOS, recognises the raid box, CentOS does not(no mapping being created in /dev).
I have also tried to create a logical RAID0 HDD on one physical HDD(as seen by MegaCLI) with MegaCLI utility in CentOS.The result was success and the new logical drive could have been used, but when restarting the server, the controllers' bios fails with an error (not surprised(one logical RAID0 on one physical HDD)) and configuration is being erased.
Has anyone tried connecting 8880EM2 controller to a raid box with raid configured on the box, running it all under CentOS and what were the results.
View 3 Replies
View Related
Sep 6, 2010
Based on the reading I've done over the past 48 hours I think I'm in serious trouble here with my RAID 5 array. I got another 1 TB drive and added to my other 3 to increase my space to 3 TB...no problem.
While the array was resyncing...it got to about 40%, I had a power failure. So I'm pretty sure it failed while it was growing the array...not the partition. Next time I booted mdadm didn't even detect the array. I fiddled around trying to get mdadm to recognize my array, but no luck.
I finally got desperate enough to just create the array again...I knew the settings of my and had seen some people have success with this method. When creating it, it asked me if I was sure because the disks appeared to belong to an array already, but I said yes. The problem is when I created it, it created a clean array and this is what I'm left with.
Code:
/dev/md0:
Version : 00.90
Creation Time : Sun Sep 5 20:01:08 2010
Raid Level : raid5
Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
[Code]....
I tried looking for backup superblock locations using e2fsck and every other tool I could find, but nothing worked. I tried testdisk which says it found my partition on /dev/md0, so I let it create the partition. Now I have a /dev/md0p1, which won't let me mount it either. What's interesting is gparted reports /dev/md0p1 as the old partition size (1.82 TB)...the data has to still be there, right?
View 3 Replies
View Related
Sep 13, 2010
on my pc, that was running WinXP, I thought of installing Ubuntu. (I did install linux already a few times in the past years and use it on another couple of pcs) But something went wrong. This machine has 2 x 200MB maxtor drives, in raid 0 configuration, supported by the motherboard Nvidia chipset, and working well in Windows. When ran the live Ubuntu 10.04 cd, gparted was not able to access the drives in raid configuration, until I installed the mdadm and kpartx packages then the existing data became visible. So after that initial moment I thought all was ok and proceeded to install Lucid on the machine, dual booting with Windows. I did partition manually so that in my 400GB raid drive there is an 80GB NTFS partition with WinXP, a 90GB extended partition for Linux Ext4 and Swap and then a last NTFS 200GB partition for data. All went well, but now on restarting the computer nothing happens, nothing loads, Grub is not showing, and it looks like I cannot launch Linux nor Windows. All the data from WinXP and the Ubuntu installation seems to be on the disks but the pc is just not booting. I suppose the problem is with the raid configuration that is not handled properly during the installation, but is there anything that I can do now, apart from reinstalling Windows Xp or installing Ubuntu in a non raid configuration?
View 9 Replies
View Related
Jun 7, 2011
I have 4 WD10EARS drives running in a RAID 5 array using MDADM.Yesterday my OS Drive failed. I have replaced this and installed a fresh copy of Ubuntu 11.04 on it. then installed MDADM, and rebooted the machine, hoping that it would automatically rebuild the array.It hasnt, when i look at the array using Disk Utility, it says that the array is not running. If i try to start the array it says :Error assembling array: mdadm exited with exit code 1: mdadm: failed to RUN_ARRAY /dev/md0: Input/output error
mdadm: Not enough devices to start the array.I have tried MDADM --assemble --scan and it gives this output:mdadm: /dev/md0 assembled from 2 drives - not enough to start the array.I know that there are 4 drives present as they are all showing, but it is only using 2 of them.I also ran MDADM -- detail /dev.md0 which gave:
root@warren-P5K-E:~# mdadm --detail /dev/md0
/dev/md0:
Version : 0.90
[code]...
View 11 Replies
View Related
May 1, 2011
I just setup sendmail on my server to send emails and it works, now I would like to be able to get an email from mdadm if sometjhing was going wrong. I imagine most raid users have this feature setup.
Right now, I have 7 raid arrays and mdadm starts at boot time. Until now, I used Mr. Goblin's script (http://connie.slackware.com/~mrgoblin/files/rc.mdadm) (thanks Mr Goblin!) to monitor my arrays.
The script is started at boot time from rc.local. I created a small script in /usr/bin that send the following command to rc.mdadm giving me the status of the arrays:
Code:
/etc/rc.d/rc.mdadm status
and it works fine, but this requires me probing the arrays manually by calling the script from the command line. I would like to automate probing every 10 minutes or whatever and if a fault has been detected, I get an email.
[Code]...
View 14 Replies
View Related
Aug 27, 2010
UPDATE: decided to reinstall and run the partitioner to get rid of the raid. Not worth dealing with this since seems to be lower level as /dev/mapper was not listing any devices. Error 15 at grub points to legacy grub. So avoiding the problem by getting rid of raid for now. So ignore this post. Found a nice grub2 explanation on the wiki but didn't help this situation since probably isn't a grub problem. Probably is a installer failure to map devices properly when it only used what was already available and didn't create them during the install. I don't know, just guessing. Had OpenSuSE 10.3 64bit installed with software raid mirrored swap, boot, root. Used the alternate 64bit Ubuntu iso for installation. Since partitioning was already correctly setup and the raid devices /dev/md0,1,2 were recognized by the installer, I chose to format the partitions with ext3 and accept the configuration:
/dev/md0 = swap = /dev/sda1, /dev/sdb1 = 2Gb
/dev/md1 = boot = /dev/sda2, /dev/sdb2 = 256Mb
/dev/md2 = root = /dev/sda3, /dev/sda3 = 20Gb
Installation process failed at the point of installing grub. It had attempted to install the bootloader on /dev/sda2 and /dev/sdb2. I moved on since it would not let me fiddle with the settings and I got the machine rebooted with the rescue option on the iso used for installing. Now, I can see the root partition is populated with files as expected. dpkg will list that linux-image-generic, headers, and linux-generic are installed with other supporting kernel packages. grub-pc is installed as well. However, the /boot partition or /dev/md1 was empty initially after the reboot. What is the procedure to get grub to install the bootloader on /dev/sda2 and /dev/sdb2, which represent /dev/md1 or /boot?
Running apt-get update and apt-get upgrade installed a newer kernel and this populated the /boot partition. Running update-grub results in a "/usr/sbin/grub-probe: error: no mapping exists for 'md2'". grub-install /dev/md2 or grub-install /dev/sda2 gives the same error as well. Both commands indicate that "Autodetection of a filesystem module failed, Please specify the module with the option '--modules' explicitly". What is the right modules that need to be loaded for a raid partition in initrd? Should I be telling grub to use the a raid module?
View 1 Replies
View Related
Jan 3, 2009
I had a RAID1 'device' build on two physical partitions on two drives. One of the disk controllers died and software RAID did the job - now I am working on the degraded array.
Now I want to put the old disk (sdb) back, and I am not sure what will happen. Both disks have 'raid auto' partitions. And sdb file structure from before of the failure. The raid code will find inconsistency between both partitions. What will it decide? Will it start coping from the currently running system (sda) all the data to the old one (sdb) at the boot time, as I wish?
I don't want to it to write from the old one to the new one, as some months passed and lots of changes happened to the data.
View 1 Replies
View Related
Jun 21, 2011
I was using Terminal and browsing a directory in my home folder. My "home" directory is located on "/dev/sdb1". When in Terminal I typed "ls" in one of my directories and the output was garbage. The output didn't show the files in the directory. I think it said something like, "input/output error". Unfortunately, I didn't write the exact error down. Instead I rebooted.The hard disk with the problem is:
Code:
$ sudo hdparm -I /dev/sdb
[sudo] password for brian:
[code]...
View 6 Replies
View Related
Dec 10, 2009
I am going to be using CentOs 5.4 for a home storage server. It will be RAID6 on 6 x 1TB drives. I plan on using an external enclosure which is connected via two SFF-8088 cables (4 drives a piece). I am looking to try and find a non-RAID HBA which would support this external enclosure and allow to use standard linux software raid.
If this is not an option, I'd consider using a hardware based raid card, but they are very expensive. The Adaptec 5085 is one option but is almost $800. If that is what I need for this thing to be solid then that is fine, I will spend the money but I am thinking that software raid may be the way to go.
View 3 Replies
View Related
Sep 14, 2010
Since I installed 11.1 and now 11.3 I haven't been abble to access my LaCie 500gb USB 2.0 external disk. So I asked LaCie support about it and they replied as follows:Windows 2k or SuSe Linux 11.0 can not see device in any USB 2.0 portsIs this a known problem and will it ever be fixed?
View 9 Replies
View Related
Mar 18, 2010
Ubuntu/Linux operating systems. It installs just fine but after it reboots I get a "disk boot failure, insert system disk". I have searched around but I can't seem to find anything that works. There is only one hard drive in my computer and no other operating systems on it.
View 1 Replies
View Related
Jan 14, 2011
While using my computer the other day (I was sending an email) it suddenly turned off. I didn't get any low power warning, but I was running on battery and had my iphone charging from a USB port.
As I didn't think there was low battery, I just turned it back on again. As it was booting I saw the battery light flashing, indicating low power. I went to get the charger, but before I got it, mid boot-up it turned off again.
This seemingly damaged something hard-disk-wise.
Upon turning it on again it dropped into busy box with some message similar to this:
Quote:
No init fount. Try passing init= bootarg.
BusyBox v1.13.3 (Ubuntu 1:1.13.3-1ubuntu7) built in shell (ash) Enter 'help' for a list of build in commands
(initramfs)
That's not the actual message (copy pasted from another post) but the message is VERY similar to that.
If I "exit" busy box, I get a load of message about "kernel panic" before it freezes up.
I have booted a live USB (what I am using now). I thought since it wasn't cleanly unmounted, simply mounting and unmounting would do the trick. I was wrong.
Code:
ubuntu@ubuntu:~$ sudo fdisk -l
Disk /dev/sda: 250.1 GB, 250059350016 bytes
255 heads, 63 sectors/track, 30401 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
[Code].....
I had previously ran e2fsck (after checking it was unmounted) but it wouldn't run also because it was reporting that the device was already mounted, busy, or being exclusively used by a process.
I don't want to do any more for fear of causing further damage.
I am astounded that such damage can be caused so easily!
View 9 Replies
View Related
Aug 21, 2011
Here are my specs to start off...
OS: Corrupted Windows XP, Home Edition
CPU: Pentium 4 2.8GHz
RAM: 1gb
Graphic Card: nVidia GeForce 8600 GS
So I corrupted my Windows OS by doing something stupid so I decided to install Unbuntu. I tried installing it over my windows OS, but as it's scanning the disk it fails and says "Apt configuration failure" and then "installation failure". I can't upload my partitions because I can't run my OS to find them.
I know the information is limited I've considered trying to install a different linux OS or even format my entire harddrive (I would really prefer NOT to do that since I'd have to install all the drivers off the internet).
View 5 Replies
View Related
Jun 10, 2010
I looking for advice on what is the best way for backup on opensuse in case hard disk failure .
I have opensuse as server running samba (with some share folder) and mysql and web service. for mysql backup I run cron job using automysqlbackup script that run 3 times a day (morning , lunch , evening).
in case of hard disk failure, I wish to be put every things on new hard disk in 1or 2 hours.
do I have to use disk imaging software ?
View 3 Replies
View Related