Ubuntu Installation :: 10.04 USB LTS: Partition #1 Failed Raid
Jul 16, 2010
The ext4 file system creation in partition #1 of Serial ATA RAID isw_fjidifbhi_Volume0 (mirror) failed. Not sure what's wrong? How do I solve this? This happens after I enter my login details and click next. Then it fails on the install screen with this error.
I get this error when trying to install 10.04 "The ext4 file system creation in partition #1 of Serial ATA RAID isw_dceecfcacg_Volume0 (stripe) failed".I have a sony Vaio with 2x256gb SSD running RAID0.
I was running Ubuntu 9.04 Desktop on a headless Pentium 4 machine which is our file, mail, web & fax server. The two x 250GB SATA hard disks were in a RAID 1 array with full disk encryption. Ran the 9.10 upgrade via WEBMIN and it failed. I should have known then to copy over everything to a backup disk, but instead I rebooted.
On restart the machine accepted my encryption passphrase but promptly hung with a mountall symbol lookup error - code 127. So I can't start the machine to get at the disks, and using a Live CD is useless as it has no way to open the RAID array to get at the encrypted partitions. Although we have data backed up (as at last night) I'd hoped not to have to rebuild the entire server from scratch. But its looking bad.I have taken one drive out and plugged it into another machine (Hercules), and the partitions show up as /dev/sdb1 /dev/sdb2 /dev/sdb3.
If it weren't for RAID, I could open /dev/sdb2 the main partition) in Disk Utility and enter my encryption passphrase to get access. But RAID adds a layer of obstruction that I have not yet overcome. I used mdadm to scan the above partitions and created the /etc/mdadm.conf file, which I edited to show the 2nd drive as missing (rather than risk corrupting both drives). I activated the RAID array with mdadm, and cat shows:
I've been searching the web for hours but have yet to find someone with a solution to this situation. If anyone has a thought on how to access this disk I'd be pleased to hear from you. In the meantime I will start building a new (9.10) machine from scratch, without RAID, 'cos that's probably going to be necessary.
I installed Fedora 12 and performed the normal updates. Now I can't reboot and get the following console error message.
ERROR: via: wrong # of devices in RAID set "via_cbcff jdief" [1/2] on /dev/sda ERROR: removing inconsistent RAID set "via_cbcff jdief" ERROR: no RAID set found No root device found Boot has failed, sleeping forever.
I am trying to install Ubuntu Server 10.04 on a home server I am making. I have 3 1-TB drives set up in RAID 5 via my mobo (ASUS M2NPV-VM). The installation detects that I have a RAID array, but when it goes to partition, all it shows me is the usb stick I am installing from.
I have 9.10 installed and working like a charm, due to the problems I had last time, I decided to leave an empty space to try and test upcoming editions, but I can't install it!
I get two error messages and can't get past the point of partition assignment.
Here's the screenshots. I'm trying to install on the free space.
I can't seem to get past step 6 of he installation of Ubuntu 10.04. I get the error: The ext4 file system creation failed... on single partition (no raid). I chose ' / ' as the mount point, and have tried with and without a swap drive. I'm installing on a Sony VAIO VGN-NS160D, and the HDD was previously formatted to NTFS. There's no other OS so I don't see any way of getting a command line to try a sudo fdisk..
Q1) I was wondering if it is possible to Dual boot Ubuntu with Windows XP on a 1TB RAID-0 setup ?
Q2) Also, is it possible to create a SWAP partition (for Ubuntu) on a NON RAID-0 HDD ?
Q3) Lastly... I read GRUB2 is the default boot manager... should I use that, or GRUB / Lio ?
I have a total of 3 HDDs on this system: -- 2x 500GB WDD HDDs (non-advanced format) ... RAID-0 setup -- 1x 320GB WDD HDD (non RAID setup) (The non RAID HDD is intended to be a SWAP drive for both XP and Ubuntu = 2 partitions)
I plan on making multiple partitions... and reserve partition space for Ubuntu (of course).
I have the latest version of the LiveCD created already.
Q4) Do I need the Alternate CD for this setup?
I plan on installing XP before Ubuntu.
This is my 1st time dual booting XP with Ubuntu.
I'm using these as my resources: - [url] - [url]
Q5) Anything else I should be aware of (possible issues during install)?
Q6) Lastly... is there anything like the AHCI (advanced host controller interface) like in Windows for Ubuntu?
(Since I need a special floppy during Windows Install...) I want to be able to use the Advanced Queuing capabilities of my SATA drives in Ubuntu.
I have now tried twice, with the same results, to install Debian 8.2.0 onto my hard drive. Each time it went through the installation as if everything was working. But, when it rebooted, it came up in grub rescue mode. When I do an "ls" command, it only shows the hard drive and no partition information. When I do an "ls (ho0,1)" it reports no such partition.
When is my installation failing? Why is it not partitioning the hard drive and loading the OS? Obviously it can write to the disk or else grub wouldn't come up at boot.
I was updating from 10.04 to 10.10 earlier today when my laptop shutoff in the middle. I fixed a grub error by reinstalling it with a live-cd, and found out I had a "kernel panic-not syncing: VFS: unable to mount root fs on" waiting for me. I booted an alternative kernel and tried fixing things from there, but got "general error mounting filesystem". So, I tried a live-cd and when I mount the partition, it is read-only. I tried running fcsk to no availThe only thing I could find helpful was this."It's probably because your filesystem has suffered a failure - it is configured by default (in /etc/fstab) to remount as read-only in such cases in order to minimise the risk of data loss."
I have installed a dozen dual boot XP / Ubuntu boxes, however this one has me stumped. After the installation failed in the partitioning phase I tried manually partitioning which also failed. Then I went to using a Gparted LiveCD. I can resize the NTFS partition fine each time rebooting in XP and it is OK. However when I shrink the NTFS then try to add an ext4 (or ext3) right after the NTFS it acts like it is working, then when it says "complete" the disk disappears out of Gparted and it shows no devices. If I reboot into the LiveCD it shows a blank second partition, if I try to format it in ext4 I get the exact same results. If I go to Windows it will show an unknown healthy partition. If I grow the NTFS back to the full 60Gb it seems to work fine. It just won't create a second primary ext4 partition.
I have a Dell Mini 9 netbook. The SSD took a dump, so I ordered a Super Talent 16gb replacement. I put it in yesterday, tried to install 10.10 and constantly get errors. The first error, which I haven't had since, was [ERROR 30] Read-only File System. The install obviously failed. Wondering if perhaps the file got messed up in translation, I redownloaded 10.10, reformatted the flash drive (sandisk cruizer 4gb), put 10.10 back on via the program on the Ubuntu site. No luck, but no longer the [error 30]. Tried again using Unetbootin, no luck. Rinse repeat a few times, no errors just a working cursor spinning and spinning and spinning and spinning and.. you get it.
Tried to put WinXP on it just because I was that frustrated, no luck.
Now I'm back to Ubuntu (because let's face it, who wants to deal with Windows, christ they make it so complicated). I'm currently using 10.04 because I was hoping (praying) it might just be a 10.10 thing.
No such luck, now it goes to step 7/7, starts, and 5% of the way thry "creating ext4 file system" it says "the ext4 file system creation in partition #1 (0,0,0)(sda) failed." I have checked the SSD in the Disk Utility, SMART tests are clean. I have gone to terminal and run fdisk and had a smarter person than me look at the copypasta, no errors, I have deleted all existing partitions in gparted and started fresh. I have tried the auto partitioning, I have tried manual, I am going insane. Literally insane. My preschooler thinks I've leaped off the deep end.
Could it be my flash drive? Could the SSD be defective despite the tests coming back clean? What do I need to type into terminal? Is there a way to entirely entirely entirely wipe the SSD to make it fresh-out-the-box clean? I will happily provide whatever you need if it means I can get my husband off my back about this stupid netbook with its stupid tiny keyboard.
I wanted to start exploring web development and perhaps hosting my own server as well as learning about linux and all the things that go with it so I downloaded the ubuntu 9.1 Server edition and burned it to a CD. I thought to put it on my Dell laptop as it is newer than my main PC and I could bring it to and fro between class. It had Vista installed and I definitely wanted to keep that in the meantime until I got more familiar with Ubuntu. The laptop has a 320GB hard drive with a 10 GB recovery partition. I went ahead and formatted the 10GB to make room for ubuntu. Also I was able to "shrink" the main windows partition by 16GB to make even more room. I could not combine the two small drives but alas. I had hoped to use the 16GB partition for the main install and the 10GB for a necessary swap drive (I am completely new to all this).
So I reboot on the server CD and get to the partition section. I was following this guide here: [url]
It seemed I did not want to do anything "guided" or "automatic" because the options were listing the entire drive and again i wanted to keep my vista untouched. So I go to manual partitioning and although the guide didn't go into enough detail I went ahead and assigned an "ext2" filetype to the larger partition and a "swap" to the smaller partition. Then I went to write changes to disk and after completing one of the two successfully the installer failed to configure the swap drive. I don't know why. I restarted to make sure windows was OK and surely it was not, as I got the dreaded "missing operating system" screen. I ran the windows recovery CD and lo and behold it could not find any drives at all, much less repair them. The data I had on the vista partition were not particularly vital, but it would be nice to have it back.
So my questions are, is there a way to recovery the windows partition? And how is the correct way to configure a dual boot system with Vista and Ubuntu 9.1 Server edition?
I grabbed the new lubuntu 10.10 from [URL] but it turns out I'm having a problem installing it on my netbook (Asus Eee PC 1015PED). While installing, this error pops up:
Quote:
The attempt to mount a file system with type ext4 in SCSI2 (0,0,0), partition #1 (sda) at / failed.You may resume partitioning from the partitioning menu.I'm installing via USB and have selected the option to erase everything and use the full HDD.
Unable to install Ubuntu 9.10 on a new internal harddrive. The hardrive contains no operating system. This hardrive is the only drive present in the system.
Whenever the installation trys to mount the ext4 partition the following error appears: The attempt to mount a file system with type ext4 in SCSI1 (0,0,0), partition #1 (sda) at /failed
Iv'e tried over and over to get past this error to no avail.
I want to change my sda2 partition to ntfs type. i have installed GParted but it is returning a strange type of error. Here is the error dump file...
[Code]...
WARNING: the kernel failed to re-read the partition table on /dev/sda (Device or resource busy). As a result, it may not reflect all of your changes until after reboot. WARNING: the kernel failed to re-read the partition table on /dev/sda (Device or resource busy). As a result, it may not reflect all of your changes until after reboot.
I have installed Ubuntu on my m1530 since 8.04 and currently dual boot Win7 and 10.10. I would like to dual boot on my PC, but I have run into a problem. I am not a pro at Ubuntu, but this problem I can not solve by reading forums like I have in the past.
I realize this is a common problem, but I have noticed people having success.
I have a M4A87TD EVO MB with two Seagate drives in Raid 0. (The raid controller is a SB850 on that MB) I use the raid utility to create the raid drive that Windows7x64 uses. I have 2 partitions and 1 unused space. Partition 1 is Windows, partition 2 is for media, and the remaining unused space is for Ubuntu.
I am running ubuntu-10.10-desktop-amd64 off a Cruzer 16GB flash drive that was installed via Universal-USB-Installer-1.8.1.4.
My problem like so many others is that when I load into Ubuntu, gparted detects two separate hard drives instead of the raid. I read that this is because kpartx is not installed on 10.10. I then went in LiveCD mode and downloaded kpartx from Synaptic Manager. Gparted still reported two drives. I opened terminal and run a few commands with kpartx. I received an error. (Forgive me I didn't write it down, but I believe it said something about a communication error. I will try again later and see.)
Currently I am reflashing the Cruzer with a persistence of 4GB. I am not familiar with this process, but I understand that my LiveCD boot will save information I download to it. I decided to try this method because I was going to install kpartx and reboot to see if this made a difference.
I am looking for any suggestions on a different method or perhaps someone to tell me that the raid controller or some hardware isn't supported. I did install ubuntu-10.10-alternate-amd64 on my flash drive, but fail to get past detecting my CD-ROM drive since it's not plugged in. If this method is viable, I will plug it in. I also watched the ..... video were a guy creates Raid 0 with the alternated CD, but it wasn't a dual boot and didn't use a raid controller from a MB.
I just had a whole 2TB Software RAID 5 blow up on me. I rebooted my server, which i hardly ever do and low and behold i loose one of my raid 5 sets. It seems like two of the disks are not showing up properly.. What i mean by that is the OS picks up the disks, but it doesnt see the partitions.
I ran smartct -l on all the drives in question and they're all in good working order.
Is there some sort of repair tool i can use to scan the busted drives (since they're available) to fix any possible errors that might be present.
Here is what the "good" drive looks like when i use sfdisk:
Quote:
sudo sfdisk -l /dev/sda Disk /dev/sda: 121601 cylinders, 255 heads, 63 sectors/track Units = cylinders of 8225280 bytes, blocks of 1024 bytes, counting from 0 Device Boot Start End #cyls #blocks Id System /dev/sda1 0+ 121600 121601- 976760001 83 Linux /dev/sda2 0 - 0 0 0 Empty
Just had my Raid 5 config on Sme server 7.4 crash due to a faulty drive, I have manage to setup a machine with ubuntu 9.04 and installed the drives in there original config (I haven't overwritten them). Ubuntu can see the raid config but for some reason I cannot mount it as it keeps asking for the filesystem which if I check says LVM2_member, anyone know how I could get it to mount?
My hard drives kept dropping out of the raid array, so I finally identified the problem as a bad sata cable. I redid the wires and now when I try to assemble I get this:
Code:
mdadm -v --assemble --scan /dev/md0 mdadm: looking for devices for /dev/md0 mdadm: no recogniseable superblock on /dev/sde4
My raid array has failed. I have two disks /dev/sda and /dev/sdb./dev/sdb has failed and I could not rebuild the array(madm returned that the device is busy) so I rebooted the machine. After that, the whole sdb disk went missing, as it now only shows sda in fdisk -l.Did the disk went totally dead or my raid glitched?
I have searched for answer to this question but cannot find one. I am trying to install ubuntu 8.04.2 on my Acer TravelMate 290 but when I get to the install stage I get the error message "the file system creation in partition 1 of sci1 (0,0,0) (sda) failed". I am dedicating the entire HD to ubuntu, no windows and no dual boot.
I'm running Karmic Server with GRUB2 on a Dell XPS 420. Everything was running fine until I changed 2 BIOS settings in an attempt to make my Virtual Box guests run faster. I turned on SpeedStep and Virtualization, rebooted, and I was slapped in the face with a grub error 15. I can't, in my wildest dreams, imagine how these two settings could cause a problem for GRUB, but they have. To make matters worse, I've set my server up to use Luks encrypted LVMs on soft-RAID. From what I can gather, it seems my only hope is to reinstall GRUB. So, I've tried to follow the Live CD instructions outlined in the following article (adding the necessary steps to mount my RAID volumes and LVMs). [URL]
If I try mounting the root lvm as 'dev/vg-root' on /mnt and the boot partition as 'dev/md0' on /mnt/boot, when I try to run the command $sudo grub-install --root-directory=/mnt/ /dev/md0, I get an errors: grub-setup: warn: Attempting to install GRUB to a partition instead of the MBR. This is a BAD idea. grub-setup: error: Embedding is not possible, but this is required when the root device is on a RAID array or LVM volume.
Somewhere in my troubleshooting, I also tried mounting the root lvm as 'dev/mapper/vg-root'. This results in the grub-install error: $sudo grub-install --root-directory=/mnt/ /dev/md0 Invalid device 'dev/md0'
Obviously, neither case fixes the problem. I've been searching and troubleshooting for several hours this evening, and I must have my system operational by Monday morning. That means if I don't have a solution by pretty early tomorrow morning...I'm screwed. A full rebuild will by my only option.
Using a fresh copy of server 10.04 im trying to simulate a failed raid array on a pair of 2tb disks. Here is the procedure i have been following so far:
- Remove the dead disk partitions from each of the raid 1 arrays (substitute the correct md devices and partitions) - mdadm /dev/md0 -r /dev/sdb2 - mdadm /dev/md1 -r /dev/sdb3
[code]....
I get an error here that sfdisk does not support gpt (guid partition table). I thought sfdisk did support gpt? It says to use parted, but i cant find a command that copies a partition table over from another disk in parted documentation. Any suggestions? I suppose i could make the partitions manually, but im writing a procedure for people who arent that technical and i need it to be simple enough to be run in my absence. manually building the partitions would be too hard for them.
Recently updated to kernel 2.5.32.11-99 and upon restart, failed to boot. Get stuck at a recovery prompt asking for root password or Ctrl-D to continue (which reboots system). I worked around this problem by reverting to the previous kernel 2.6.32.9-70. From what I can tell, with the newer kernel there is a problem with device enumeration of software raid volumes.
I have a home server running Openfiler 2.3 x64 with 4x1.5TB software RAID 5 array (more details on the hardware and OS later). All was working well for two years until several weeks ago, the array failed with two faulty disks at the same time. Well, those thing could happen, especially if one is using desktop-grade disks instead of enterprise-grade ones (way too expensive for a home server). Since is was most likely a false positive, I've reassembled the array:
Code:
# mdadm --assemble --force /dev/md0 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 mdadm: forcing event count in /dev/sdb1(0) from 110 upto 122 mdadm: forcing event count in /dev/sdc1(1) from 110 upto 122
[code]....
Right. Once is just a coincident but twice in such a sort period of time means that something is wrong. I've reassembled the array and again, all the files were intact. But now was the time to think seriously about backing up my array, so I've ordered a 2TB external disk and in the meantime kept the server off. When I got the external drive, I hooked it up to my Windows desktop, turned on the server and started copying the files. After about 10 minutes two drives failed again. I've reassembled, rebooted and started copying again, but after a few MBs, the copy process reported a problem - the files were unavailable. A few retried and the process resumed, but a few MBs later it had to stop again, for the same reason. Several more stops like those and two disks failed again. Looking at the /var/log/messages file, I found a lot of error like these:
Quote:
Apr 12 22:44:02 NAS kernel: [77047.467686] ata1.00: configured for UDMA/33 Apr 12 22:44:02 NAS kernel: [77047.523714] ata1.01: configured for UDMA/133 Apr 12 22:44:02 NAS kernel: [77047.523727] ata1: EH complete
[code]....
The motherboard is Gigabyte GA-G31M-ES2L based on Intel's G31 chipset, the 4 disks are Seagate 7200.11 (with a version of a firmware that doesn't cause frequent data corruption).
Could any RAID gurus kindly assist me on the following RAID-5 issue?I have an mdadm-created RAID5 array consisting of 4 discs. One of the discs was dropping out, so I decided to replace it. Somehow, this went terribly wrong and I succeeded in marking two of the drives as faulty, and the re-adding them as spare.
Now the array is (logically) no longer able to start:
mdadm: Not enough devices to start the array.Degraded and can't create RAID ,auto stop RAID [md1]
Code: mdadm --create --assume-clean --level=5 --raid-devices=4 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2 As I don't want to ruin the maybe small chance I have left to rescue my data, I would like to hear the input of this wise community.
I recognize that this isn't the typical question, but I have a problem with my OpenSUSE webserver, and I thought I would prevail on the community for some guidance. I have this webserver with an important MySQL db on it. The RAID array seems to have died while I was moving. (did someone drop it? dunno) Now it can't find any boot device. It has 4 old SCSI drives.
So, I know how to mount a IDE or SATA drive as a slave, in a Linux environment to read data off of it (to copy the MySQL files off of it.) But, how do I do that with a SCSI drive? Also, I have an additional (identical) server to the crippled one. What will happen if I just slide one of the scsi drives into the operating server? Is this second identical server going to help me at all? I don't even know what is on it. Can I reconfigure the RAID, so it's not using a drive, and then slide in a disk from the crippled server, and copy the data off of it?