Ubuntu :: Upgrade System While Using Software RAID?

Jun 25, 2010

Is it possible to upgrade an Ubuntu install which is using software RAID?

I ask because I'm currently using Fedora, and one of my huge frustrations with the distro is that if you're using software RAID, the installer explicitly notes you can't upgrade to future versions. Instead, it requires that you wipe your entire drive and do a full reinstall to acquire the next release. I presume Ubuntu probably uses many of the same libraries to support software RAID as Fedora, so does Ubuntu also have this limitation?

View 3 Replies


ADVERTISEMENT

Fedora Installation :: F12 Won't Boot On LVM/RAID System After Upgrade

Dec 9, 2009

I upgraded from F10 to F12 using preupgrade. The upgrade itself completed with no errors, but I'm unable to boot afterward.

Symptoms:...Grub starts, the initramfs loads, and the system begins to boot. After a few seconds I get error messages for buffer i/o errors on blocks 0-3 on certain dm devices (usually dm0 and dm2, but I can't get a shell to figure out what those are). An error appears from device-mapper that it couldn't read an LVM snapshot's metadata. I get the message to press "I" for interactive startup. UDEV loads and the system tries to mount all filesystems. Errors appear stating that it couldn't mount various LVM partitions. Startup fails due to the mount failure, the system reboots, and the steps repeat.

Troubleshooting done:...I have tried to run preupgrade again (the entry is still in my grub.conf file). The upgrade environment boots, but it fails to find the LVM devices and gives me a question to name my machine just like for a fresh install. I also tried booting from the full install DVD, but I get the same effect. Suspecting that the XFS drivers weren't being included, I have run dracut to create a new initramfs, making sure the XFS module was included. I have loaded the preupgrade environment and stopped at the initial GUI splash screen to get to a shell prompt. From there I can successfully assemble the raid arrays, activate the volume group, and mount all volumes -- all my data is still intact (yay!). I've run lvdisplay to check the LVM volumes, and most (all?) appear to have different UUIDs than what was in /etc/fstab before the upgrade -- not sure if preupgrade or a new LVM package somehow changed the UUIDs. I have modified my root partition's /etc/fstab to try calling the LVM volumes by name instead of UUID, but the problem persists (I also make sure to update the initramfs as well). From the device-mapper and I/O errors above, I suspect that either RAID or LVM aren't starting up properly, especially since prior OS upgrades had problems recognizing RAID/LVM combinations (it happened so regularly that I wrote a script so I could do a mkinitrd with the proper options running under SystemRescueCD with each upgrade).

I have tried booting with combinations of the rootfstype, rdinfo, rdshell, and rdinitdebug parameters, but the error happens so early in the startup process that the messages quickly scroll by and I just end up rebooting.

System details:4 1-TB drives set up in two RAID 1 pairs. FAT32 /boot partition RAIDed on the first drive pair. Two LVM partitions -- one RAIDed on the second drive pair and one on the remainder of the first drive pair. Root and other filesystems are in LVM; most (including /) are formatted in XFS.

I've made some progress in diagnosing the issue. The failure is happening because the third RAID array (md2) isn't being assembled at startup. That array contains the second physical volume in the LVM volume group, so if it doesn't start then several mount points can't be found.

The RAID array is listed in my /etc/mdadm.conf file and identified by its UUID but the Fedora 12 installer won't detect it by default. Booting the DVD in rescue mode does allow the filesystems to be detected and mounted, but the RAID device is set to be /dev/md127 instead of /dev/md2.

The arrays are on an MSI P35 motherboard (Intel ICH9R SATA chipset) but I'm using LInux software RAID. The motherboard is configured for AHCI only. This all worked correctly in Fedora 10.

View 3 Replies View Related

Server :: Convert RAID 1 To NON-RAID System In Cent OS 5 System?

Jan 6, 2010

Here is my system: I have dell poweredge 1950 PERC 6 with 300 GB raid system. It has two disks of each 300GB RAID mirrored system. I have few applications and data that reached around 280GB. As you know, poweredge 1950 we can have only two disk.

They are not mission critical. Hence, I wanted to remove the raid system and use as a non-raid system. By doing it, The applications and data can grow upto 600GB. I do not want to loose the data and setup. I am not so clear about RAID system and its conversion.

View 6 Replies View Related

General :: Migrate An Installed Ubuntu System From A Software Raid To A Hardware Raid?

Jun 29, 2011

migrate an installed Ubuntu system from a software raid to a hardware raid on the same machine? how would you go about doing so?

View 1 Replies View Related

Ubuntu :: Converting RAID 1 To RAID 5 System

Jun 12, 2010

I have two 1TB hard drives in a RAID 1 (mirroring) array. I would like to add a third 1TB drive and create a RAID 5 with the 3 drives for a 2TB system. I have ubuntu installed on a separate drive. Is it possible to convert my RAID 1 system to a RAID 5 without losing the data? Is there a better solution?

View 1 Replies View Related

Server :: Tur A Non-RAID System To RAID?

Mar 24, 2011

I have a box that doesn't have a Raid controller or a software raid running currently. I would like to make it a RAID 1. Since it seems there isn't any IDE RAID controllers hardly around, I have another HD that is the exact model as the drive currently in the box running CentOS. Can I some how add the second drive and get the box to mirror from here own out? The box gets really hot and I want to be ready for a HD failure.

View 3 Replies View Related

General :: Recover Raid - LVM On Non System Drives After System Failure

Jan 19, 2011

I have (had) Debian Testing running on a 250GB IDE hard drive, partitioned normally.

I also have 4x 1TB drives in a raid 5 using mdadm, and 2x 500GB drives in a raid 1 also with mdadm.

I put the two arrays in lvm using:

I then used "lvcreate" to make storage/backup 300GB, and the rest went to storage/media (approx. 2TB usable). I put an xfs filesystem on both and mounted them.

All was working fine until the system drive shorted out and died on me this morning. As far as I can tell, all my other drives and everything else is fine. I do a daily rsnapshot of the filesystem, which of course is residing on storage/backup (stupid, I know). So I have full backups of everything, but I'll have to put a new hard drive in and reinstall Debian before I can restore everything.

I've reinstalled before and simply reassembled mdadm arrays and remounted them before with no problems, but this is the first time I've used lvm, so I'm not sure what I have to do to restore everything. Is it as simple as reinstalling the system then doing a:

View 4 Replies View Related

Ubuntu Servers :: RAID-5 Recovery (spare/active) / Degraded And Can't Create Raid ,auto Stop Raid [md1]?

Feb 1, 2011

Could any RAID gurus kindly assist me on the following RAID-5 issue?I have an mdadm-created RAID5 array consisting of 4 discs. One of the discs was dropping out, so I decided to replace it. Somehow, this went terribly wrong and I succeeded in marking two of the drives as faulty, and the re-adding them as spare.

Now the array is (logically) no longer able to start:

mdadm: Not enough devices to start the array.Degraded and can't create RAID ,auto stop RAID [md1]

I was able to examine the disks though:

Code:
root@127.0.0.1:/etc# mdadm --examine /dev/sdb2
/dev/sdb2:
Magic : a92b4efc
Version : 00.90.00
code....

Code:
mdadm --create --assume-clean --level=5 --raid-devices=4 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2
As I don't want to ruin the maybe small chance I have left to rescue my data, I would like to hear the input of this wise community.

View 4 Replies View Related

Ubuntu Installation :: Raid Array Not Available After Upgrade?

Jul 11, 2010

After upgrading my ubuntu install my raid array is gone. The drives appear in blkid as "Linux raid member" and both have the same uuid. If I try to mount the drive via fstab I get a message that the drive is not ready or present. If I try to mount each of the two drives, one mounts successfully the other reports serious errors. Issuing a cat /proc/mdstat shows md_d0 as inactive.How can I re-establish my raid array? I have the data backed up so if I have to wipe out the disks to start over that's an option.

View 2 Replies View Related

Ubuntu Servers :: Upgrade 8.04 To RAID 1 After Install / DMRAID Necessary?

Nov 27, 2010

I have two Ubuntu 8.04 servers currently. I've purchased some 500 GB hard drives and a RAID cage just yesterday and plan to change my current servers to using RAID1 with the hard drives. Much like everyone else, I have the nVidia RAID built onto the motherboard that I plan to use. There are many how-to's out there on how to setup RAID using dmraid during the install process - but what would a person do if they are simply changing over to a RAId system? I installed dmraid a few days ago on the server. It seems that even though I have RAID shut off in the BIOS, it saw the one and only drive as a mirrored array. When I rebooted today, the server would not start; it had ERROR: degraded disks.. bla bla. Then Grub tried to start the partition based off the UUID of the drive (which was not changed) and it said "device or resource busy" and would not boot.

This problem was corrected by going into the BIOS and turning on RAID. When I rebooted, the nVidia RAID firmware started and said degraded. I went in there and deleted the so-called mirror keeping the data intact. Rebooted, disabled the RAID feature in the BIOS and then the server loaded normally with a message "NO RAID DISKS" - but at lest it did boot! So this leads me to believe that trying to turn on a RAID-1 array (just mirror between two disks) may be a challenge since I already have the system installed. I will be using Partimage to make an image of the current hard drive and then restore it to one of the new drives.

The next question is - is it a requirement that dmraid is even installed? While I understand that the nVidia RAID is fake-raid - if I go into the nVidia RAID controller and setup the mirroring between the two disks before I even restore the files to the new disk, will both drives be mirrored? However, I do think that I'd probably have to have dmraid installed to check the integrity of the array. So, I'm just a bit lost on this subject. I definitely do not want to start this server over from scratch - so curious to know if anyone has any guidance on simply making a mirrored array after the install procedure, what kind of Grub changes will be needed (because of the previous resource is busy message), and whether installing dmraid is even a requirement.

View 9 Replies View Related

Ubuntu Installation :: Desperate: Upgrade From 10.4 64-bit To 10.10 - Won't Boot From Raid

Dec 2, 2010

I upgraded my Dell T3500 from lucid64 to maverick64. The system cannot boot the current image becuase it gives up waiting for root device. it gives something like:

Code: ---udevd-work[175]: '/sbin/modprobe -bv pci: v00008086d00002822sv00001028s00000293bc01sc04i00' unexpected exit with status 0x0009 ALRT! /dev/mapper/isw_bbfcjjefaj_ARRAY01 does not exist. dropping to a shell! at which point it drops to busybox at the (initramfs) prompt. I cannot boot this kernel in either normal or recovery mode (ubuntu 10.10, Kernel 2.6.35-23-generic)

however, if I select the previous kernel in grub, (2.6.32-26) it will boot fine. This is my main production machine and it hosts my Project Management software for my team.

View 9 Replies View Related

Ubuntu Servers :: Hardware Upgrade And Existing RAID?

Apr 20, 2011

I have an old server I build 7 years ago. I have a RAID5 with 5x2TB HD's and I'm planning a major hardware overhaul. My server currently runs on a Pentium4 3.2 Ghz (pre multicore technology) on a SuperMicro mobo. I'm planning to switch to an AMD Phenom II X6 1100T Black Edition Thuban 3.3GHz, 3.7GHz Turbo 6.

So here's the question. Can I just plug my drives to the new board and restart the RAID like nothing happened? I don't have space to backup all my data if I have to recreate the RAID from scratch.

View 3 Replies View Related

Ubuntu Installation :: GRUB2 / RAID 10.10 To 11.04 Upgrade Fail

Apr 30, 2011

GRUB2 / RAID 10.10 to 11.04 Upgrade Fail.....

View 9 Replies View Related

Ubuntu :: Hardware RAID Showing As Seperate Partitions After 10.04 Upgrade?

Sep 27, 2010

I have been trying to upgrade my server to Ubuntu 10.04 since it has come out, but I hit a roadblock with my hardware RAID I have two JBODs that work perfectly in Ubuntu 9.10 x64 - but show as seperate, unformatted partitions (one per hdd) in Ubuntu 10.04 x64. Here's the relevant portion of my fstab:

UUID="1fac4a2c-6a2e-4278-bab4-c7179c8720ee" /media/Motherload ext4 exec,auto,async,rw 0 2 #4TB-ext4-/dev/mapper/pdc_cabedbfaef1
UUID="31f1f023-e6a0-4b5e-b9d9-0a46011e7fa0" /media/Fry ext4 exec,auto,async,rw 0 2 #3TB-

[code]....

View 9 Replies View Related

Ubuntu Installation :: Upgrade To 9.10 Failed; Can't Boot And Have Encrypted RAID Disk?

Feb 23, 2010

I was running Ubuntu 9.04 Desktop on a headless Pentium 4 machine which is our file, mail, web & fax server. The two x 250GB SATA hard disks were in a RAID 1 array with full disk encryption. Ran the 9.10 upgrade via WEBMIN and it failed. I should have known then to copy over everything to a backup disk, but instead I rebooted.

On restart the machine accepted my encryption passphrase but promptly hung with a mountall symbol lookup error - code 127. So I can't start the machine to get at the disks, and using a Live CD is useless as it has no way to open the RAID array to get at the encrypted partitions. Although we have data backed up (as at last night) I'd hoped not to have to rebuild the entire server from scratch. But its looking bad.I have taken one drive out and plugged it into another machine (Hercules), and the partitions show up as /dev/sdb1 /dev/sdb2 /dev/sdb3.

If it weren't for RAID, I could open /dev/sdb2 the main partition) in Disk Utility and enter my encryption passphrase to get access. But RAID adds a layer of obstruction that I have not yet overcome. I used mdadm to scan the above partitions and created the /etc/mdadm.conf file, which I edited to show the 2nd drive as missing (rather than risk corrupting both drives). I activated the RAID array with mdadm, and cat shows:

Code:
root@HERCULES# cat /proc/mdstat
Personalities : [raid1]
md1 : active raid1 sdb3[0]
1815232 blocks [2/1] [U_]

[Code]...

I've been searching the web for hours but have yet to find someone with a solution to this situation. If anyone has a thought on how to access this disk I'd be pleased to hear from you. In the meantime I will start building a new (9.10) machine from scratch, without RAID, 'cos that's probably going to be necessary.

View 1 Replies View Related

Ubuntu Installation :: Kernel Upgrade To 2.6.32-23-generic Broke My LVM - RAID And GRUB

Jul 4, 2010

I recently updated the kernel to the newest version(2.6.32-23-generic) using the update manager and now I am unable to boot in to my Lucid installation.

My setup is LVM on top of a RAID 0 array. My computer had been running in this configuration since Lucid was released.

The Error Code I get on Boot is to the effect of: /dev/mapper/ubuntu-os is not available.. and then I get kicked in to Busy Box.

Once in Busy Box if I try to use mdadm to mount the RAID array I get this error:

If I boot in to the live CD I can mount all of the partitions and LVM volumes, so it does not appear to be a failed drive or volume.

I have looked in the mdadm.conf, lvm config and grub config files and searched the "Google" for an answer with no-avail...

Ultimately I would like to find a solution which doesn't involve a re-installation.

View 6 Replies View Related

Server :: RAID 1 Partition Devices Disappear After Upgrade

Mar 3, 2011

I found a workaround of sorts. It looks like this is related to a 9.04 bug [URL] and the loopback workaround brings back the array. It is not clear how I will handle this long term.

Note: before using this technique, I used gparted to tag the partitions as "raid". They disappeared again on reboot, so I had to do it again. I am not sure how this is going to work out long-term

Note: I suspect some of this is related to the embedded "HOMEHOST" that is written into the RAID metadata on the paritions. The server was misnamed when first built and the name was changed later (cerebus -> cerberus) and the old name has surfaced in the name of a phanton device reported by gparted - /dev/mapper/jmicron_cerebus_root

I have a mythbuntu 9.10 system that I have upgraded from 8.10 to 9.04 to 9.10 in the last 2 days. I am on my way to 10.x, but need to make sure it works after every step.

The basic problem is that in its current incarnation, it is not recognizing the underlying partitions for one of the RAID devices, and therefore not happy.

As a 8.10 system I had 2 raid devices:

/dev/md16 -> /dev/sda5 and /dev/sdb5
/dev/md21 -> /dev/sdc1 and /dev/sdd1
/etc/fstab looked like this (in part):
/dev/md16 /var/lib xfs defaults 0 2

[Code]....

I don't *really* want to repartition the drive as there is a small amount of data loss between recent backups and what is on the drive, plus it would take me 2 days to move the data back.

View 2 Replies View Related

Ubuntu :: System Automatically RAID Itself?

Sep 6, 2010

When I installed this system (Xubuntu 9.04 x64) several months ago, I had two identical SATA hard drives, but I didn't do a RAID1 mirror then because I didn't want to wipe out my old OS (FreeBSD) on the second drive in case I needed to retrieve something from it. So I installed Xubuntu on the first drive (sda) and for all that time, the second drive (sdb) has been running but unused, and fdisk showed that the FreeBSD partition was still there.

A couple weeks ago, my separate backup server failed, so as a short-term backup I did a 'dd if=/dev/sda of=/dev/sdb' to make an exact copy of my Xubuntu install on drive A upon drive B (the former FreeBSD drive), then mounted /dev/sdb1 on /mnt to make sure it worked successfully. So far so good, but I still didn't set up any RAID stuff.

Several days later, I needed to reboot because of a security upgrade, and when I logged out, the GUI froze up. Thinking not too much of it, I restarted the system, and it came up fine. But the next day, I discovered that I was missing several days worth of email and recent files. In fact, everything had been reverted back to a state from several days earlier --- I think from the day I copied the first drive to the second one. Files I created in the interim were gone, and files I deleted in the interim were back. It was as if the first drive was 'restored' from the second one without my knowledge.

Doing some testing now, I find that if I create a new file on /dev/sda1 and then mount /dev/sdb1, the file also exists on sdb1. It's as if they're acting as a RAID1 mirror, without my telling it to do so. Could it just decide to do RAID1 because it sees there are two identically partitioned drives? That seems dangerous. And if they were really in a RAID mirror, why would it let me mount them separately? It's very strange.

I don't mind if it's suddenly decided to do RAID, but I want to make sure it's not going to 'restore' a more current filesystem from an older one again, if that's indeed what happened.

View 3 Replies View Related

Ubuntu :: RAID Data Drive For System?

Apr 7, 2010

I am trying to create a RAID data drive for my system but I am having setting it up since I am a total linux noob.

The system has 3 physical HDD-s:
1. 320 GB (has functional Ubuntu 9.10 installation) attached to a PCI SATA card
2. 2TB on motherboard
3. 2TB on PCI SATA card

I want to create a software RAID1 of disks 2 and 3. So far I have used the Palimpsest Disk Utility:
- Created a GUID Partition table on both disks (2, 3)
- Used File -> New -> Software Array, made sure both my drives were included
- Once Palimpsest listed the RAID Drive as a Software RAID Array, I told it to create Ext3 filesystem on it

Well.. at least thats what I thought I did. At this point I have been able to mount the RAID drive and put files on it. However when I look at its information in Palimpsest, I am told that the drive is not partitioned. Both RAId components /dev/sda1 and /dev/sdc1 are reported to be in Sync, but the RAID Drive's own state is 'Running, Resyncing @ 45%' (and lowly growing).

My questions are: Is this a normal setup or did I do something incorrectly? Why is the drive reporting to have no partition? And howcome I can use it if it does not have a partition? I have found the command line based configurations to be a tad too confusing to follow, so I have tried to stick to graphical tools - is this a hopeless cause in Ubuntu or is it possible to achieve what I want to do without command line? I will list some info on my disks below - perhaps this offers more insight to those of you more familiar with Linux.


Code:
mindgamer@mind-server:~$ sudo lshw -C disk
[sudo] password for mindgamer:
*-disk:0
description: ATA Disk
product: WDC WD3200BEVT-0
vendor: Western Digital

[Code]...

View 7 Replies View Related

Ubuntu :: Hardware RAID System Very Slow

May 2, 2010

As the title says, my raid system is very very slow (this is for raid5 6x1Tb SAMSUNG HD103UJ):
Code:
leo@server:~$ sudo hdparm -tT /dev/sdb
/dev/sdb:
Timing cached reads: 1094 MB in 2.00 seconds = 546.79 MB/sec
Timing buffered disk reads: 8 MB in 3.16 seconds = 2.53 MB/sec
It's impossible and I've tried every configuration (raid5,raid0,Pass Through), and I become nearly exactly the same speed (with restarts, so I'm sure I'm really talking with the volumes I've defined).

One can compare it with the system drive :
Code:
leo@server:~$ sudo hdparm -tT /dev/sda
/dev/sda:
Timing cached reads: 1756 MB in 2.00 seconds = 878.12 MB/sec
Timing buffered disk reads: 226 MB in 3.01 seconds = 75.14 MB/sec

Some hardware/software infos :
Code:
Raid card
Controller Name ARC-1230
Firmware Version V1.48 2009-12-31

Code:
Motherboard
Manufacturer: ASUSTeK Computer INC.
Product Name: P5LD2-VM DH

Code:
leo@server:~$ uname -a
Linux server 2.6.31-20-server #58-Ubuntu SMP Fri Mar 12 05:40:05 UTC 2010 x86_64 GNU/Linux

Code:
IDE Channels .....
I'm a bit lost now. I could change the motherboard, or some bios settings.

View 7 Replies View Related

Ubuntu Installation :: Can't Install 10.04 On Raid 1 System?

Jun 22, 2010

I tried to install new ubuntu on Intel raid 1 system but it said that:

Quote:

The ext4 file system creation in partition #1 of Serial ATA RAID isw_chibcceegh_Volume0 (mirror) failed.

My config is:
P5Q Pro
2x500 GB Seagate HDD
Intel Raid 1

Boot ubuntu from USB Drive (Wonder does this cause the problem?)

View 9 Replies View Related

Ubuntu :: Encrypt System Drive - RAID 1

Feb 22, 2011

1- I want to encrypt my system drive
2 - I want to make my system drive a RAID 1
3 - I have two 1 T hard drives, but I would like to be able to add two more 1 T hard drives for a total of 4 1T drives in the future. How do I do that?

I am using ubuntu 10, desktop edition. I have not added the OS yet, I am building this weekend.

View 2 Replies View Related

Ubuntu Installation :: Upgrade From Another Distro To Server - Migrating Existing LVM - No Encryption And No Software Raid

Mar 28, 2010

I want to upgrade from another distro to ubuntu server for a few reasons. The only problem is I have a lot of data that needs to survive. here is how my computer is setup. I've 5 drives on the computer,

A- 10gb drive for OS and swap only, no data

B,C,D,E - 4x 500 GB drives in a LVM. they make up one large drive with xfs and this volume has about 1.2 TB of data. there is nothing fancy on it, no encryption and no software raid of course the little 10gb drive can be formatted no problem, but the LVM needs to be migrated over intact.

View 5 Replies View Related

Ubuntu :: Want To Upgrade \ How The System Will Look At Things To Upgrade

Nov 7, 2010

I decided to give Ubuntu another try and installed the Studio version. During my installation it asked whether I wanted the Realtime kernel or not. I chose yes as I figured it could come in handy when I work with audio.However since this version isn't exactly Ubuntu 10.10 as promoted , I decided to install Ubuntu Desktop on top of it (trough apt-get). For the full experience. ( im not even sure if it just installed extra software and added some folders)Now my questions are:1:The Realtime kernel is it now overwritten by any other kernel that came with the Desktop installation?

2: How will the updates handle my installation once i want to upgrade to, lets say, 11.04.I'm not really sure about how the system will look at things to upgrade. Will it look at my Desktop installation or my Ubuntu Studio? ( concerning the Realtime kernel as well).Just to add some more: I'm a total Linux Noob(although im not scared to use the Terminal)

View 3 Replies View Related

General :: Raid - Recover Software RAID5 Array After Server Upgrade?

Jun 7, 2011

I recently upgraded a server from Fedora 6 to Fedora 14. In addition to the main hard drive where the OS is installed, I have 3 1TB hard drives configured for RAID5 (via software). After the upgrade, I noticed one of the hard drives had been removed from the raid array. I tried to add it back with mdadm --add, but it just put it in as a spare. I figured I'd get back to it later.Then, when performing a reboot, the system could not mount the raid array at all. I removed it from the fstab so I could boot the system, and now I'm trying to get the raid array back up.

I ran the following:mdadm --create /dev/md0 --assume-clean --level=5 --chunk=64 --raid-devices=3 missing /dev/sdc1 /dev/sdd1I know my chunk size is 64k, and "missing" is for the drive that got kicked out of the array (/dev/sdb1).That seemed to work, and mdadm reports that the array is running "clean, degraded" with the missing drive.However, I can't mount the raid array. When I try:mount -t ext3 /dev/md0 /mnt/fooI get:

mount: wrong fs type, bad option, bad superblock on /dev/md0,
missing codepage or helper program, or other error
In some cases useful info is found in syslog - try

[code]....

View 1 Replies View Related

Ubuntu Installation :: Possible To Have A Dual Boot System And Have It In RAID 0 ?

Mar 21, 2010

Is it humanly possible to have a dual boot system and have it in RAID 0

I have three Hard drives.....
The three drives that I have are250GB IDE
1TB Sata
1TB Sata

And I edit . produce a lot of music/videos

What I am looking to do it have my 250Gb IDE drive as my Operating system (partitioned in two or something, one for Windows Another for Ubuntu)

And I would like when I am on either Operating System (ubuntu/windows_7)
its see's ,my computer as having a 2TB HDD (setting the two 1TB HDD's in RAID 0 )

So in Laymans... 250GB for OS's and 2TB (2 x 1TB HDD) as DATA drives

View 1 Replies View Related

Ubuntu Servers :: Moving Raid System To A New Host?

Aug 3, 2011

A while back I successfully set up a software raid (mdadm) install of ubuntu server on a cheap Compaq machine. I'm pretty sure it's running an nForce 430 chipset. Anyway, I put 3 2TB drives in it and set up two arrays: a small raid1 array for /boot and a raid5 array occupying the rest of the drives.

One point of note here is that the raid1 array never seemed to "take hold": the partition is empty, and the system has been booting entirely from the raid5 array. I'm not sure if this is relevant or not: just throwing it in there in case it is.

Anyway, I recently upgraded my desktop PC, and planned to hand down the Q6600 and DG33TL motherboard from it to the server. I did the hardware upgrade, but now the system will not boot. My question is: why.

Boot fails with the following error message:

"No bootable device -- insert boot disk and press any key"

This to me suggests that it's not even getting to the point of running grub2: the mobo itself just isn't finding anything it considers to be bootable.

I've tried all combinations I can think of of configuring the SATA support in the BIOS (modes include IDE, ACHI and RAID, all in combination with UEFI boot enabled and disabled). The system does detect the three drives: it lists them in the BIOS config screen.

I've booted from an ubuntu rescue USB drive and I'm able to see the drives and even assemble the array. I can mount the LVM partitions therein and do what I'd normally be able to do with them.

Everything *seems* fine: it just can't boot.

Grub2 is something of a mystery to me, so I'm not exactly sure what I need to do to get it booting again, if indeed it is a problem with grub.

View 9 Replies View Related

Server :: Ubuntu 10.04 LTS - RAID 1 Base System Installation

Dec 30, 2010

I'm trying to install a box with Ubuntu 10.04 LTS server with a typical LAMP system in order to replace my old Ubuntu 8.04 server I have at my school to have a "backup" system and also trying to replace NIS authentication with LDAP. Well I'm getting stuck on the first step: installation of base system. I want to build a RAID-1 system with the two 320GB SATA HD's the machine has, I have a little experience in installing RAID because I installed my old 8.04 server with RAID-1 aswell. I boot my box from an USB stick with Ubuntu 10.04 Server 64bit, the systems boots well, asks me about language, keyboard and so, finds the two NIC cards and the DHCP configuration of one of the card is done.

Then it starts the partitioner. One of the HD's already contains three partitions with an installation of a regional flavour of Ubuntu, the other HD only contains a partition for backups, I don't want to preserve all this stuff, so the first thing I do is to replace the partition tables of both HD's with a new one. This is done without problems. Then I go to the first disk, by example, and create a new partition, the partitioner ask me for the size, I write 0.5 GB (or 500 MB), then I select that it has to be a Primary partition at the beginning of the disk, all goes OK.

Once created I go to the "Use as: " line, type Enter and select the option "Partition for RAID volume", when I hit Enter the error appears: the screen flickers in black for a second or two, then it shows a progress bar "Starting the partitioner..." That always get stuck at 47%!!! Sometimes the partitioner allows me progress a little further (for ex. lets me activate the boot bit of the partition, or it allows me to make another partiton, even once the error didn't appeared until the first partition of the second HD!!) but it always get stuck with the same progress bar at the 47%.

I've tried a lot of things: I downloaded again the ISO and rebuilt the USB, same result. Downloaded the ISO and rebuilt the USB from another computer, same result. Unplugged all the SATA and IDE drives except the two HD's, same result. Built a CD-ROM instead of an USB, same result. Downloaded the 10.10 server ISO (not an LTS), and the USB stick can't boot, is another error, but only to try.

When the error appears, I hit Ctrl+Alt+F2 and get into a root prompt, there I kill two processes: /bin/partman and /lib...don'tremember/init.d/35... and then when I return to the first console the progress bar has gone and the install process asks me at wich step I want to return, I hit "Partition disk" and then the progress bar reappears and Stucks immediately at 47%!!! Is the Ubuntu 10.04 LTS server installer wrong?

View 5 Replies View Related

Ubuntu Installation :: No Root File System Is Defined - 10.04 / RAID

May 3, 2010

I am running a RAID0 array, with Windows 7 Ultimate x64 installed.

When i install LL10.04 through Wubi, it installs fine, reboots, continues the installation procedure, then it gives me an error box "No root file system is defined".

I have attempted pressing the "OK" button 10 or 15 times, however it does not progress. The box just keeps on popping up. My only option is a hard reset.

I've tried downloading the latest version of Wubi from the official website, and allowing Wubi to download ubuntu itself, and still nothing.

I do not want to create a new partition for Ubuntu and use the GRUB loader. I have a multi boot system and would like to stick to the windows boot loader.

View 9 Replies View Related

Debian :: System Rescue With Hardware RAID

Mar 1, 2016

I had a RAID problem on an HP Proliant server due to a failing disk. When I changed the disk, things got complicated and RAID seemed to be broken. I put back the old disk, repaired the RAID, then changed put the new disk again and all returned to normal except the system doesn't boot. I am stuck at grub stage (the grub rescue prompt). I grabbed a netinst CD and tried to rescue it, at some point, the wizard correctly sees my two partitions sda1 and sda2 and asks wether I want to chroot to sda1 or sda2. I had red screen errors on both.

The error message said to check syslog, syslog says can't mount EXT4 file system beacuse bad superblock. I switched to TTY2 (Alt-F2) and tried fsck.ext4 on sda1 (I think sda2 is the swap because when I ran fsck on it it said something like this partition is too small and suggested that it could be swap) it says bad superblock and bad magic number. I tried e2fsck -b 8193 as suggested by the error message but that too didn't work (I think the -b 8193 is for trying to get he backup superblock).

The RAID is as follows : One RAID array of 4 physical disks that are grouped into one Logical Volume /dev/sda, so the operating system only sees one device instead of four (4 disks).

View 0 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved