Hardware :: Installing New RAID Card Causes Existing Arrays To Have Bad Magic Number
May 13, 2010
I have two SAS RAID controller cards in a Dell server in slots 2 & 3, both with an array hanging off them. I went to install a third card into slot 1, but then when it boots it says two of my sd's have bad magic number in the super-block and it wants me to create an alternative one, which I don't want to do. If i remove the new card, the server boots perfectly like it did before I added the new card. Is the new card trying to control stuff that isn't hooked up to it because its in slot 1, so its confusing RHEL?
I'm having some (well, a lot actually) of problems trying to get OpenSUSE 11.2 installed on my home PC. I am attempting to set up a dual boot configuration with Windows 7 installed on an bios Nvidia RAID 0.I was able to shrink the partition in Windows, and rebooted onto the net install for OpenSUSE (the MD5 validated DVD install failed multiple burns with "Repository not found"). I get into the graphical installer portion with no problems off the net install CD. However, the installer is not recognizing that there is an existing RAID 0. It lists the 2 SATA disks in the RAID separately. I can click on SDA1, and both SDB and SDB1 and it shows the disks, but does not recognize any existing partitions. If I click on SDA I get an immediate segfault in YaST and drop back to the text mode installer menu. It is loading the nv_sata module just fine.
From forum searches and google it seems that this is not usually a problem. My motherboard is a Gigabyte GA-M57SLI-S4 with the Nvidia Nforce 570 chipset running an AMD X2 64 3800+. Removing the stripeset and redoing it as Linux software RAID is not an option, I do not have enough space for a total backup/restore. Anything I do has to be nondestructive to the existing Windows installation.I really want to have a linux installation but between the DVD installer failing and now this issue I am about ready to give up on it.
After lots of trouble with the new Nvidia stuff in Fedora 12 I gave up and am trying to switch to an ATI Radeon 4350 card in the existing system.I have removed the Nvidia refs with rpm -e, deleted /etc/X11/xorg*, and installed the card. I am only able to get a maximum resolution of 1024x768. Here's what I've done:
Boot and su/delete all refs: /etc/X11/xorg* Reboot. System comes up with 1024x768 Select Administration->Display to run system-config-display. Su/password dialogue
[code]....
If I try to add a mode line following Depth 24, It is ignored by system-config-display and overwritten. This worked with the Nvidia installation and I was able to select my resolution above 1024x768.
I've been setting up a number of netbooks for friends in a group I belong to. I am installing Ubuntu 10.04 and/or Linux Mint 9 on all of the netbooks. The netbooks are all identical. Everything has seemed OK until yesterday. I tried to set the label on the new disk using: # e2label /dev/sda newlabel The error is Bad magic number in super-block while trying to open /dev/sda. Couldn't find valid filesystem superblock.
But the filesystem seems fine. It is mounted. It booted up without errors. But all the e2* tools give the same error. So I checked another netbook with an identical fresh install of Ubuntu 10.04 32-bit. It gives the same error although the system appears to work fine. Next I went and checked my own netbook. It is identical except I swapped out the HDD for an OCZ Vertex SSD. I aligned the partitions and prepared that drive carefully. It seems to work great. But when I ran e2label on it I got the same error message.
All the other installs are fresh, clean, standard installs where I let the installer use the whole disk and do the partitioning automatically. All are using ext4. I did a fresh install of Linux Mint 9 and it has the same error as Ubuntu 10.04. Kubuntu 10.04 on my desktop gives the same error too.
This has to be the sixth time I am going to have to reinstall Ubuntu (Can't get enough of that Linux greatness) in the past 6 months. I just don't understand what exactly is wrong, I don't install anything extra (but the ubuntu restricted extras) yet updates that ask me to reboot will take me to a GRUB command line screen and trying to boot through the command line interface gives me an, "Invalid Magic Number", error.
There should be an official tool for when problems like this come up, so that users like me can easily move their files off their HDD before reinstalling. I didn't come here to rant but, does anyone know what may be causing this everytime, how it can be avoided maybe even how to deal with the Magic Number error and boot in (I wouldn't mind typing in the commands every time BTW)?
My system Hp Compaq 6730s 2.0Ghz 575 Celeron 4GB RAM Intel X4500MHD Broadcom 802.11b/g
I've compiled and added a kernel in Gentoo before. It doesn't seem to go quite as smoothly in Kubuntu 9.10 These are the steps I followed: I unpacked the kernel in /usr/src and ran make && make modules_install succesfully. Then I copied the kernel in arch/x86/boot/bzImage to /boot/bzImage-2.6.32 This entry is the one given by Kubuntu:
Code:
menuentry "Ubuntu, Linux 2.6.31-14-generic" { recordfail=1 if [ -n ${have_grubenv} ]; then save_env recordfail; fi set quiet=1
[code]....
I just read the script that update grub uses. Changed the name of the kernel from xyz to vmlinuz-2.6.32-generic and it worked.
With mdadm was only able to add a new drive to the using the --force function. I do not feel comfortable with using the function that way though. When I remove a disk in VMWare, it perfectly says that the drive is lost and the array is degraded (mdadm --detail /dev/md0). Although after re-adding the drive, it immediately shows device as busy for both mdadm and sfdisk when I don't use --force.
Recovery and repairation of degraded array worked fine with sfdisk --force, mdadm --add --force, it automatically started recovering and took not so long. What are best practises to manage software raid-1 arrays?
Have a customer who is due for a new system. AS they just renewed their RHEL entitlement, they plan on ordering a Dell server without a OS preload. Two questions:
- Will RH let them download RHEL6 just by maintaining the entitlement when their current version is RHEL 3?
- The server will have two RAID arrays - one intended for /home, one for "everything else". As I've never done a clean load with two arrays, how do I select what file systems go on which array?
I am in a situation where I am stuck with a LVM cleanup process. Although I know a lot about AIX LVM , but this is first time I am working with Linux LVM2. Problem is that I created two RAID arrays on storage, which appeared as mpath0 & mpath1 devices (multipath) on RHEL. I created logical volumes and volume groups and every thing was fine till I decided to clean the storage arrays and ran following script:
#!/bin/sh cat /scripts/numbers | while read numbers do lvremove -f /dev/vg$numbers/lv_vg$numbers vgremove -f vg$numbers pvremove -f /dev/mapper/mpath$numbersp1 done
Please note that numbers was a file in same directory, having numbers 1 and 2 in separate line. Scripts worked well and i was able to delete definitions properly (however I now think I missed one parted command to remove the partition definition from mpath device. When I created three new arrays, I got devices from mpath2 to mpath5 on linux and then I created vg0 to vg2. By mistake, I ran above script again for cleanup purpose and now I got following error message
Cant remove physical volume /dev/mapper/mpath2p1 of volume group vg0 without ff[/B]
Now after doing mind search, I now realize that I have messed up (particularly because mpath devices did not map in sequence to vg devices and mapping was like mpath2 --- to ---- vg0 and onwards). Now how I can cleanup the lvm definitions? should i go for pvremove -ff flag or investigate further? I am not concerned about data, I just want to cleanup these pv/vg/lv/mpath definations so that lvm can be cleaned up properly and I can start over with new raid arrays from storage?
I'm trying to setup a RAID 5 array of 3x2TB drives and noticed that, besides having a faulty drive listed, I keep getting what looks like two separate arrays defined. I've setup the array using the following : sudo mdadm --create /dev/md01 --verbose --chunk=64 --level=5 --raid-devices=3 /dev/sdb /dev/sdc /dev/sde
So I've defined it as md01, or so I think. However, looking in the Disk Utility the array is listed as md1 (degraded) instead. Sure enough I get :cat /proc/mdstat Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10] md1 : active raid5 sde[3](F) sdc[1] sdb[0] 3907028992 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]
So I tried getting info from mdadm on both md01 and md1 :user@al9000:~$ sudo mdadm --detail /dev/md1 /dev/md1: Version : 00.90 Creation Time : Sun Jan 9 10:51:21 2011 Raid Level : raid5 ......
Is this normal? I've tried using mdadm to --stop then --remove both arrays and then start from scratch but I end up in the same place. I'm just getting my feet wet with this so perhaps I'm missing some fundamentals here. I think the drive fault is a separate issue, strange since the Disk Utility says the drive is healthy and I'm running the self test now. Perhaps a bad cable is my next check...
I have created software raid 5 configurations on the second harddrive its working fine and i have edited fstab file for auto mounting when it reboot but when i reboot the computer raid doesn't work i have to re-create the arrays by typing "mdadm --create" command again and mount again manually ,is there anywhere i can do this once without retyping the commands again after rebooting and i am also using redhat 5
what do I have:2x 150GB drives (sda) on a raid card (raid 1)for the OS (slack 13.37)2x 2TB drives (sdb) on that same raid card (raid 1, too)2x 1.5TB drives (sdc,sdd) directly attached to MoBo2x 750GB drives (sde,sdf) attached to MoBo too.if i got about it the normal way, i'd create softRAID 1 out of the the 1.5TB and the 750GB drives and LVM all the data arrays (2TB+1.5TB+750GB) to get a unified disk.If I use btrfs will I be able to do the same? mean I have read how to create raid arrays with mkfs.btrfs and that some lvm capability is incorporated in the filesystem. but will it understand what I want it to do, if i just say
I have never preformed a rebuild of an RAID array. I am collecting resources, which details how to build an RAID 5 array when one drive has failed. Does the BIOS on the RAID controller card start to rebuild the data on the new drive once it is installed?
With mdadm was only able to add a new drive to the using the --force function. I do not feel comfortable with using the function that way though.
When I remove a disk in VMWare, it perfectly says that the drive is lost and the array is degraded (mdadm --detail /dev/md0). Although after re-adding the drive, it immediately shows device as busy for both mdadm and sfdisk when I don't use --force.
Recovery and repairation of degraded array worked fine with sfdisk --force, mdadm --add --force, it automatically started recovering and took not so long.
What are best practises to manage software raid-1 arrays?
So I have a system that is about 6 years old running Redhat 7.2 that is supporting a very old app that cannot be replaced at the moment. The jbod has 7 Raid1 arrays in it, 6 of which are for database storage and another for the OS storage. We've recently run into some bad slowdowns and drive failures causing nearly a week in downtime. Apparently none of the people involved, including the so-called hardware experts could really shed any light on the matter. Out of curiosity I ran iostat one day for a while and saw numbers similar to below:
[Code]...
Some of these kinda weird me out, especially the disk utilization and the corresponding low data transfer. I'm not a disk IO expert so if there are any gurus out there willing to help explain what it is I'm seeing here. As a side note, the system is back up and running it just runs sluggish and neither the database folks nor the hardware guys can make heads or tails of it. Ive sent them the same graphs from iostat but so far no response.
I'm working on a midrange NAS system basically running on Linux and I got to do some great testing today. The step-by-step lead me to using fsdb to corrupt the magic number on a file system in order to corrupt it / test the script that should fix it.
After several crashes during videos it seemed like a good idea to fsck root. Downloaded the latest systemrescuecd and ran it at boot. The error message was 'bad magic number, corrupt superblock' with a suggested command to try another superblock. That failed with the same message. Tried tune2fs to force fsck at boot and got the same message. The drive is less than 6 months old and the installed system is working more or less ok. The command I used was 'fsck.ext4 /dev/sdc2'. What am I doing wrong?
i wanted to see the red hat side of things and do some virtualization with CentOS, so i am trying to dual boot ubuntu 10.04 LTS and CentOS 5.5. the machine is a laptop, toshiba A100 series. what I did was to create the following partitioning scheme via Ubuntu LiveCd
Code: Device Boot Start End Blocks Id System /dev/sda1 * 1 19457 156288290+ 5 Extended /dev/sda5 1 2103 16892284+ 83 Linux /dev/sda6 2104 9988 63336231 83 Linux /dev/sda7 9995 12623 21117411 83 Linux /dev/sda8 12624 18800 49616721 83 Linux /dev/sda9 18801 19457 5277321 82 Linux swap / Solaris
created an extended partition and in there have made sda5 and sda6 as / and home for ubuntu and sda7 and sda8 as / and home for CentOS. and sda9 as swap. I installed ubuntu first and then installed CentOS with no bootloader. Run sudo update-grub through ubuntu and now i have both Ubuntu and CentOS available. But when i select CentOS, i have an error which reads "invalid magic number".
I have grub2 installed, haven't downgraded or done anything to it and the ubuntu install is fresh, one week since i updated to 10.04 from scratch. I have found much contradictory stuff on google, but not something that provides a definite solution and also this post but the second command provided in the solution is one i cannot understand very well and it doesn't seem to work. what I am doing wrong here and how to make this work. I would prefer to do things via Ubuntu since debian stuff is what i am comfortable with and i am installing CentOS to learn not to do work on it.
I tried the steps usually offered (modifying the grub.cfg file, replacing the wubildr files), it still wouldn't boot.
I booted in with the LiveCD to try to chroot into the environment and try running 'update-grub'. However, whenever I tried to chroot, it said it could not execute /bin/bash!
Trying to modify the grub commands manually (which originally gave the error: you need to load the kernal first), in the end said "invalid magic number"
There are some configuration things set in this installation I would like to not lose. Is it hopeless?
I loaded all required files inside the Innovate ARM board.When It starts booting from that location it is showing a error "Bad magic number" How I can get out of this error.
How long does hardware Raid card (raid 1, 2 drives)take to mirror a 1 TB drive (500gb used)?Is there a general rule of thumb for this?4 hours? 12 hours? 24 hours?
Currently I have 3 hard drives 2pcs 10gb almost the same 1pc 20gb
I have a layout of (10.2gb) /dev/hda1 boot 104391 83 Linux /dev/hda2 9912105 8e Linux LVM (10.1gb) /dev/hdb1 9873328+ 8e Linux LVM (20.4gb) (unpartitioned)
The two 10g is setup as lvm and I want to make raid1 using the 20gb hdd. Almost all I see is raid1 first in the internet.
I have a system that has the following partitions:
Now SDC is a new drive I added. I would like to pool that new drive with the raided drives to give myself more space on my existing system (and structure). Is this possible since my raid already has data on it?
I have a RAID5 with 5x2TB HD's and I'm planning a major hardware overhaul. My server currently runs on a Pentium4 3.2 Ghz (pre multicore technology) on a SuperMicro mobo. I'm planning to switch to an AMD Phenom II X6 1100T Black Edition Thuban 3.3GHz, 3.7GHz Turbo 6.
So here's the question. Can I just plug my drives to the new board and restart the RAID like nothing happened? I don't have space to backup all my data if I have to recreate the RAID from scratch. Intel and AMD should be binary compatible (I mean the RPM's work) so I should be able to invoke mdadm to assemble the RAID after I install Linux on the new server. Right?