CentOS 5 :: Installing 5.4 With Multiple Raid Levels On A 4 Disk Array?
Nov 17, 2009
Our server is a CybertronPC I2XV9080 Imperium Tower. It is equipped with a supermicro X7DVL-I Motherboard and Quad 750 GB SATA2 RAID edition hard drives in a raid 5 array. We tried to install Centos on the Raid5 array with Device-Mapper as the LVM. In the BIOS SATA Raid was enabled and the ICH RAID code base option was set to [Intel].
Intel Matrix Storage Manager Option ROM V5.6.4.1002 ESB2
RAID
ID Name Level Strip Size Status Bootable
0 Raid5 Raid 5 64KB 80GB Normal Yes
1 Raid_5 Raid 5 64kB 2000GB Normal Yes[code].....
Can I have multiple level raids across the same array or would that lead to problems as above? Is the root cause of my problem the fact that intel raid5 is not supported for Linux as based on the following link http:[url]....
View 3 Replies
ADVERTISEMENT
Sep 27, 2010
I have a NETGEAR ReadyNAS NV+ with four 1TB drives in a RAID-5 array. This is our primary file storage. This has previously been backed up to a hardware RAID-0 array directly attached to our Windows server. The capacity of this backup array is no longer sufficient. So the plan was, take a bunch of 200GB to 320GB drives (And a 750) I had kicking around, chuck them in a couple of old SCSI drive enclosures I have collecting dust, attach them via IDA/SATA-to-USB adaptors to a USB hub, attach that to the server, create a JBOD array spanning the disks, and back up the NAS to that. Performance is not an issue as this is just to be used for backup, with the idea being as near to zero cost as possible (Spend so far = NZ$100�ish).
The first hurdle I struck was Windows not supporting Dynamic Disks on USB drives (Required to create a spanned volume). At first I resisted using another machine (i.e. a machine running Ubuntu) as I didn't want to dedicate a piece of hardware to backing up the NAS. I then decided it would be acceptable to do this via a VM, which is what I've done.So I have 10.04 running under VMWare Server 2.0.2 under Windows Server 2008 R2. The disks are all presented to the VM. I wasn't sure if I was going to end up creating the array under LVM or something else, but I noticed Disk Utility has an option to create an array, so I tried that. When I add two 250GB drives, the array size is 500GB. When I then add a 160GB drive, the array size drops to 480GB. Huh? If I keep adding disks (Regardless of order) the final array size comes out at 1.8 TB, as per the attached screenshot. Now with the following drives, I expected something more like:
160 + 250 + 250+ 750 + 250 +200 + 200 + 250 + 320 + 250 + 320 = 3.2TB
Am I missing something or making a false assumption somewhere?
View 4 Replies
View Related
Aug 1, 2010
I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.
I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).
These are the commands I used:
Quote:
p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*
[Code]....
I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...
So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.
It's still trying to use /dev/disk/by-uuid/412d... to boot.
What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.
View 2 Replies
View Related
May 23, 2011
I need to set up a RAID 1 array on Squeeze. I have 3 partitions: sda1 is root, sda5 is home, and sda6 is swap. (sda2 is the extended partition containing home and swap. This was a clean installation, so I don't know what happened to sda3 and sda4...)
All the information that I've been able to find recommends doing something like this:
mdadm --create --verbose /dev/md0 --level=mirror --raid-devices=2 /dev/sdb1 /dev/sdc1
Do I need to type a separate command for each partition, or is there a better way to do it? Also, should I use the UUID instead of the dev names?
View 4 Replies
View Related
Dec 2, 2010
Alright, I have this issue on both SystemRescueCD and Debian Squeeze. I have an ASUS P5Q Turbo board that supports hardware RAID. If I configure an array and then start the Linux installer or boot the rescue CD, I get /dev/sda and /dev/sdb instead of an array. What gives? I need to start installing within the hour so I am desperate for an answer!
View 1 Replies
View Related
Aug 31, 2010
concerning Linux, mdadm, and creating RAID Array's in Debian. I've done a lot of reading and research on RAID both on this board and elsewhere (The Linux Documentation Project's Software-RAID HOWTO is especially good), but I've run across something that no one seems to explain, and I'm not sure why. I'm instructed to create partitions on the drives I wish to add to my array. These partitions inevitably take up the whole disk, and are always have their system IDs set to "Linux raid autodetect". What I don't understand is why, after creating these partitions, some guides then go on to create an array (say a RAID5 one) with just the disks themselves as members, while others go on to create the RAID5 array with the previously created partitions as members. E.g.,
mdadm --create --verbose /dev/md0 --level=5 --raid-devices=4 /dev/sda /dev/sdb /dev/sdc /dev/sdd
vs.
mdadm --create --verbose /dev/md0 --level=5 --raid-devices=4 /dev/sda1 /dev/sdb1 /dev/sdc1 /dev/sdd1
What's the advantage of using one over the other?
View 3 Replies
View Related
Jul 8, 2010
What is the best way to install Windows and Linux on two-hard-disk array? In fakeraid there are no problems in Win, but linux installation is almost impossible (i've tried unsuccessfully...). In software raid it would be impossible to share files between win and linux? And finally hardware raid is possible, but cheap controllers have low performance. Is there any other way (apart from spending a lot of $$ for adaptec controller) ?
View 1 Replies
View Related
Dec 19, 2010
I went to setup my linux box and found that the OS drive had finally died. It was an extremely old WD raptor drive in a hot box full of drives so it was really only a matter of time before it just quit on me. Normally this wouldn't be such a big deal however I had just recently constructed an md RAID5 array of 3 1TB disks to act as an NFS mount for basically all of my important files. Maybe 2-3 weeks before the failure I had finished moving all of my most important stuff onto that array. Now I know that the array is intact. All the required data is sitting on those disks. Since only the OS level disk failed on me I should be able to get a new disk in there, reinstall ubuntu and then rebuild that array. how exactly do I go about doing that with mdadm? Do I create the array from the /dev character devices like when I initially built the array?
View 2 Replies
View Related
Oct 30, 2010
I installed RC1, and never touched the disks that were raided previously, and the attached image is what i see for my raid array that was previously working fine.
View 3 Replies
View Related
Oct 27, 2010
We have some servers that run in very harsh environments (research vessel) that need to have high-availability.We have software RAID 1 for some measure of resiliency, along with proper data backups (tapes etc), however we would like to be able to break out a new server and re-image it (including RAID setup) from a known good copy if the hardware completely fails on the production box. Simplicity of the process is a big plus.I am interested in any advice on the best way to approach this. My current approach (relatively new to Linux administration, totally new to MDADM) is to use DD to take a complete gzipped copy of one of the RAID'ed devices (from a live CD): ode:
dd if=/dev/sda bs=4096 | gzip -c > /mnt/external/image/test.img then reverse the process on the new PC, finally using Code:mdadm --assemble to re-create and re-build the array.
View 1 Replies
View Related
May 11, 2011
Round two:I am trying to install a RAID 1 array on my system. I already have another RAID 1 array in there. I am using the BIOS RAID option to set up the array.Here's what dmraid -r tells me:
# dmraid -r
/dev/sda: pdc, "pdc_bajfedfacg", mirror, ok, 3906249984 sectors, data@ 0
/dev/sdb: pdc, "pdc_bajfedfacg", mirror, ok, 3906249984 sectors, data@ 0
[code]....
View 9 Replies
View Related
Jul 20, 2009
I'm not able to spin down my harddisks in raid array. I did hdparm -S 240 /dev/sdb and the same for the second hdd. Smaller values like 2 minutes work. I don't have log files on this disks. How can I determine, which processes are accessing these disks? I have apache and samba running. Nobody is accessing the server.
View 2 Replies
View Related
Mar 10, 2010
I have a raid 1 array which I did setup on a Centos box. I can configure the array fine the problem is every time is restart my machine I can no longer see the array and have to go create it all over again. I tried doing a few searches on it but have came up with nothing so far.
View 19 Replies
View Related
Jul 15, 2010
I've been having troubles with software raid. In particular, the raid array becomes un "assembleable" after reboots. The config is CentOS 5, 4 sata discs (one by 160 containing OS, no raid and 3 2TB disks configured as a RAID 5 array - no spare drive). These drives were configured in anaconda and all seemed to go well (the drive and its lvm partitions worked and it finished rebuilding overnight). A couple of reboots later the drives cannot be assembled anymore and the machine won't boot. The error message says:
mdadm: /dev/md0 assembled from 1 drive and 1 spare - not enough to start the array.
Of course there are 3 drives and no spares in the array as configured. Manually starting the array with mdadm --assemble --scan gives the same message as does assembling the drive by specifying the individual parts. /proc/mdstat does recognize the 3 drives and when I look at the partition tables in fdisk, they show as being software raid. What could be wrong or steps to diagnose? I tried configuring the raid drives manually before going the anaconda route. Also, does anyone know I can edit the /etc/fstab file to disable them so the machine will at least boot. The (Repair filesystem) shell has the / drive mounted r/o.
View 7 Replies
View Related
Jun 23, 2009
I need to mount my raid array on CentOS 5.2 samba server.
Here are my hardware specs:
Motherboard: Tyan S2510 LE dual PIII
CPU's: Intel PIII 850ghz socket 370
Memory: 4 gig Crucial 133 ECC SDRAM
OS: 2 x'x IBM Travelstar 6.4 gig 2.5 hard drives, (low heat/noise)
Storage: 4 x's Seagate 500 gig IDE 7200 rpm
RAID controller: 3Ware 7500-12 controller, (RAID 5) (66 mhz PCI bus)
NIC: 3COM 3C996B-T gigabit NIC, (66 mhz PCI bus)
I have the 2 IBM's set as RAID 1, (mirror) and the 4 Seagates as RAID 5, (1.5 TB) I have installed the OS with minor problems, (motherboard doesn't like the 2.6.18-128.1.14.el5 kernel, removed it from my grub.conf).
My problem is mounting the RAID array. I have done the following:
formatted with fdisk;
fdisk /dev/sdb
Then formatted with the following command;
mkfs.ext3 -m 0 /dev/sdb
The hard drive was formatted with the ext3 files system, but I have mounted it as an ext2 file system as I don't want 'journaling' to occur. I then edited my /ect/fstab like this: .....
Then: mount -a
When I go into my "home" directory and type ls, I get the following:
[root@hydra home]# ls -l
total 24
drwx------ 2 zog zog 4096 Jun 23 15:50 zog
lrwxrwxrwx 1 root root 6 Jun 23 15:46 home -> /home/
drwxrwxrwx 2 root root 16384 Jun 23 15:34 lost+found
drwxr-xr-x 2 root root 4096 Jun 23 17:18 tmp
Why my home directory is showing under home?
View 5 Replies
View Related
Sep 23, 2011
I have a home samba server with a 3ware Escalade 8506-8. I have 5 x 500 gig hard drives in a RAID 5 array. Recently, my 8506 died and I need to get a new one. However, I saw a 3ware Escalade 9500S-12 on ebay for about $20.00 dollars more than a replacement 8506-8.
My question is, if I put my drives on the 9500S, will it recognize my existing RAID array? Or will it want to build a new RAID array and format all of my data?Hope I have asked this question clearly, little short on sleep this week.
View 3 Replies
View Related
Mar 25, 2010
I have a 10x2tb disk array that i'm trying to build into a single software raid 5 i have tried this 2 times now the first it made it to 58.7% and the machine locked up and the array would not restart after a reboot. On my 2nd try all was looking good until about 50% i noticed that the speed dropped in 1/2 and that ksoftirqd/2 is taking up a lot of cpu (about 90%) the md0_resync and md0_raid5 are also taking 60-90% when the build started they took 7%. when i do a dmesg i see a lot of the message compute_blocknr: map not correct.
For a little info on the physical setup this is running on an Atom 510 with 2GB of mem the drives are connected to an addonics 4-Port RAID 5 / JBOD SATA II PCI Controller using the sil3124 chipset. I'm using 2 addonics 5X1 SATA Port Multiplier connected to the controller to get the 10 drives attached. All drive show up and don't seem to have any issues. I'm running a fully updated as of 3/20/10 version of centos 5.4
I will let this continue to run over night but i expect it to be locked up by morning if it follows what the last attempt did.
View 9 Replies
View Related
Dec 12, 2010
how to rename disk array
# df -kh
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/VolGroup00-LogVol00
444G 4.4G 417G 2% /
[code]....
View 3 Replies
View Related
May 5, 2011
I have a couple of Centos servers, each connected to the Internet using its Static IP address. They are in the same physical rack. Is there a way I an get them to share a disk array. The disk array could be on one of the servers, or it could be separate SAN array.
View 3 Replies
View Related
May 6, 2009
I have an old server based on Tyan Thunder LE-T (s2518UGN),: embeeded VGA - ATI Rage XL two Intel Pentium III (PGA370) processors at 1133 MHz 512 MB ECC RAM embeeded Dual Intel 82559 LAN controllers 10/100Mbps embeeded Adaptec 7899W NT Ultra160 SCSI Adaptec SCSI RAID 2000S (PCI card to support for bootable RAID 5) 3x Seagate HDD ST318438LW connected in RAID 5 instaled OS was RedHat 8 rack case (2U) It worked fine but one disk from array failed.
I provided backup and decided use two mirrored SATA disks Seagate ST3500320NS HDD ES.2 500GB with PCI SATA controler Kouwell KW-5125 (SiI3124 chipset) inserted through PCI riser card (2U case). (I can't replace old SCSI disk because it isn't available any more.) I disabled SCSI controler in BIOS and removed Adaptec SCSI RAID 2000S PCI card. I inserted PCI SATA controler connected both SATA disks and configured mirroring.
[Code]....
View 10 Replies
View Related
Apr 28, 2009
I've just setup a new centos 5.3 server with a 3-disc raid 5 software RAID array. I've setup other software raid 5 arrays on this same hardware, whilst testing, and had no trouble... I only just installed 3 new drives and performed a new install from scratch.
Hardware is: 4800XP X2 64-bit, 2GB RAM, Albatron KI-690G mainboard with Marvel SATA controller (I think) - 4 ports.
SATA port 0 is system drive (OCZ vertex 32GB SSD)
SATA port 1 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 2 is Western Digital "green" 1TB 8MB cache SATA 2
SATA port 3 is Western Digital "green" 1TB 8MB cache SATA 2
Ports 1-3 in software raid MD0 (raid 5)
All of this was configured in the GUI setup, with LVM on top of software RAID 5 mounted on /var. Partition size is ~1.9TB
Trouble is I get all sorts of "hardware failure" messages at bootup and the MD driver reports it was only able to bring up 2 out of 3 drives in the RAID set... however the RAID set formatted fine during the setup?
Here is the relevent dmesg output...
View 2 Replies
View Related
Oct 23, 2009
hHow would I add my SSH server to run levels 3,4,5? I have been trying to search google for this answer to my problem for hours and cannot have any luck.
View 2 Replies
View Related
Feb 15, 2009
The installer can't see my raid controller (I assume) as I'm getting the following error:"Error opening /dev/mapper/isw_jbhgjgjj_Vol0: No such device or address"It just sees them as 4 individual drives: sda, sdb, sdc and sdd.Please note that I have set up the RAID 5 in the controller bios interface and the image name is Vol0, which it seems that it tries to load but for some particular reason it can't.I have also tried different bios settings and nothing worked.
View 3 Replies
View Related
Jan 13, 2010
My problem is that I'm trying to install CentOS 5.4 x86_64 DVD ISO on Supermicro X7SBI server with installed Adaptec RAID 3405 controller.
I created RAID 5 array and is working fine (adaptec status says Optimal) but I can't install CentOS to that array (1.5TB size).
Whenever I try to install with: linux dd
I'm asked for a driver, which I have downloaded from Adaptec site and extracted contents to USB drive (in installation found as /sba1) which has now a lot of IMG and some ISO files on it.
I try to load (I simplified names) RHEL5.img, CENTOS.img... with x64 names (one exact name: aacraid driverdisk-CentOS-x86_64.img) and I always get the error message: "No devices of the appropriate type were found on this driver disk"
This is going on for a week now and I can't find the right driver or something I'm doing wrong to get install done.
View 16 Replies
View Related
May 7, 2009
I need your help urgently. I will shortly install Centos 5 on a HP DL380R05 E5420 Server with two HP 146GB 10k 2.5 SAS HP SP hard disks running hardware RAID-1. I am a newbie on CentOS, please advise me:
1. Do I need a Centos RAID device driver for hardware RAID-1 on the HP DL380R05 E5420 Server?
2. Should I use Centos i386 or Centos x86_64 in this case? Which one: CentOS 5.2 or CentOS 5.3?
3. Does Centos support Embedded NC373i Multifunction Gigabit network cards?
The following is my server's hardware configuration:
HP DL380R05 E5420 2GB Base AP Server(1) Quad-Core Intel Xeon Processor E5420 Two Embedded NC373i Multifunction Gigabit / 2 GB (2 x 1 GB) PC2-5300 Fully Buffered DIMMs (DDR2-667)/HP Smart Array P400/256MB Controller /Rack (2U)/Hot Plug Fully Redundant Fans Standard Two HP 146GB 10k 2.5 SAS HP SP HDD HP 1000W IEC C14 Cord RPS Kit
View 19 Replies
View Related
Dec 15, 2010
So I didn't notice when I setup my CentOS 5.5 server that I left / as RAID 0 on md1. All the rest are RAID 1. Is there a way I can modify the array to RAID 1 without a risk of data loss? I'm glad I caught this before I setup any other services. I've only setup smb so far...
[root@ftpserver ~]# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/md1 16G 3.0G 13G 20% /
[code]....
View 1 Replies
View Related
Dec 23, 2010
I have a RAID 5 array, md0, with three full-disk (non-partitioned) members, sdb, sdc, and sdd. My computer will hang during the AHCI BIOS if AHCI is enabled instead of IDE, if these drives are plugged in. I believe it may be because I'm using the whole disk, and the AHCI BIOS expects an MBR to be on the drive (I don't know why it would care).
Is there a way to convert the array to use members sdb1, sdc1 and sdd1, partitioned MBR with 0xFD RAID partitions?
View 1 Replies
View Related
Jun 9, 2011
Following scenario: My server in some data center on a different continent with two disks and software raid 1.
One day I see that a disk failed (for example with /proc/mdstat). Of course I should replace the failed disk asap. Now that I think about it, I am not sure how. What should my email to the data center support guy mention to make sure that guy doesn't replace the wrong disk?
With hardware RAID it is very easy, because the controller usually has some kind of red LED indicator. But what about software raid?
View 8 Replies
View Related
Feb 5, 2010
I am setting up a samba server to operate in a windows AD domain. I want to set permissions for multiple groups to have different levels of access to one group of files, and it looks to me like unix permissions will not do that? I always hear about how robust linux is, and it seems to me that their file permissions model is WEAK compared to microsoft's?
View 2 Replies
View Related
Mar 9, 2009
I tried install Centos 5.2 (i386) on Dell PowerEge 2950 server with SAS RAID disk (RAID 5 on PERC 6/i created from six HDD, 1 TB each). I never installed Centos on a disk over 2 TB.
I find out that i need use gpt instead mbr. I tried use rescue mode + parted but after rebooting I got message "Can't boot from from GPT partition" whet tried install. But when the space is partitioned by installer, I can't boot up.I got msg PBR not found. My ask is: "Can I boot directly from GPT partition? If yes how do it?" Or even better "How to install centos 5.2 on disk bigger than 2 TB?"
View 2 Replies
View Related