Ubuntu :: 10.10 - Automounting A Software RAID Device?
Dec 29, 2010
Just set up a home server with Ubuntu 10.10 desktop. I've set up a software RAID 1 device using System/Administration/Disk Utility which seems to work well. However when I reboot the machine and try and access the drive I get the error that 'authentication is required to start this raid device' and then I have to enter my password, after which all is good
View 2 Replies
ADVERTISEMENT
Dec 30, 2010
Just set up a home server with Ubuntu 10.10 desktop. I've set up a software RAID 1 device using System/Administration/Disk Utility which seems to work well. However when I reboot the machine and try and access the drive I get the error that 'authentication is required to start this raid device' and then I have to enter my password, after which all is good.
View 1 Replies
View Related
Aug 22, 2010
I am running a RAID 6 array via mdadm and I cannot make it autostart and automount at boot which is fairly imperative for a server config.I am running the latest build of regular Ubuntu and not server, mostly because I have some other tasks for it that kind of requires a GUI, amongst that virtualization that I have so far working correctly.Can anyone provide me with proper instructions on how to get the array running properly?I found a few guides, added some things to the mdadm.conf which was supposed to be the end of it, however the array still doesn't start and let alone automount.
View 9 Replies
View Related
Sep 15, 2010
It's been a real battle, but I am getting close.I won't go into all the details of the fight that I have had, but I've almost made it to the finish line. Here is the set up. ASUS Z8PE-D18 mother board 2 CPU, 8 Gig Ram. I recently added an OCZ Agility SSD, defined a raid 1 virtual disk on the 1 terabyte WD HDD drives, which will holds all of my user data, the SSD is for executables.The bios is set to AHCI. Windows 7 installed fine, recognizes the raid VD just fine.
I installed Ubuntu 10.04 by first booting into try and mode, then opening a terminal and issuing a "sudo dmraid -ay" command. Then performing the install. I told it to install the raid components, and told it to let me specify the partitions manually. When setting up the partitions, I told it to use the free space I set aside on the SSD from the Windows 7 install as ext4 and to mount root there. Ubuntu installed just fine, grub2 comes up just fine, and Windows 7 boots with out a hitch, recognizing the mirrored partition as I indicated previously. When I tell grub to boot linux however, it pauses and I get the "no block devices found" message. It will then boot, but it does not recognize the raid array. After Ubuntu starts up I can run "dmraid -ay" and it recognizes the raid array, but shows the two component disks of the raid array as well. It will not allow the component disks to be mounted, but they show up which is annoying. (I can live with that if I have to)
I have fixed a similar problem before by setting up a dmraid script in /etc/initramfs-tools/scripts/local-top ... following the instructions found at the bottom of this blog:[URL].. To recap: My problem is that after grub2 fires up Ubuntu 10.04.1 LTS (Lucid Lynx), it pauses, and I get "no block devices found" It then boots but does not recognize the raid array untill I manually run "dmraid -ay". I've hunted around for what to do but I have not found anything. It may be some timing issue or something, but I am so tired of beating my head against this wall.
View 4 Replies
View Related
Jan 16, 2010
I have three WD 1.5 GB harddrives. 2 of them already in a linear RAID also called Concatenated i think. (the same as JBOD). Can i add the third drive to the RAID without losing data? Update "Using mdadm software raid."
View 7 Replies
View Related
Dec 27, 2010
let's say this system has 3 hard drives. Drive #1 and #2 are RAID 0 and Windows7 lives there. It is a hardware RAID, not software.
On Drive #3 Ubuntu has been installed using WUBI - it boots up and works okay - but it does not see the RAID array.
Do I just need a linux driver to be able to see & mount my "Windows" RAID0 array? Or is this even possible? Can anyone point me in the right direction?
View 1 Replies
View Related
Jan 29, 2011
I've decided to toy around with LVM and mdadm this weekend. I can get everything working, and all is well, until I restart. After that, I no longer have any /dev/md0 device, which during the auto mount process, causes an error. I've looked through several HOWTOs, as well as the LVM/mdadm man pages, and I believe I've tracked it down to mdadm's "assemble" that is needed (so that LVM can see the md0 device).
Not exactly sure how to go about having this occur during the boot process to ensure that the LVM mapped drive is available for when fstab is read. In case it helps this is a base install of 10.10 server 64. I have four drives, the first is used for the OS and is not in the RAID array (nor LVM). The second and third are RAID1 (/dev/md0) and there is a volume group associated with /dev/md0. The last is a LVM, but not RAID, and it has its own volume group.
View 9 Replies
View Related
Feb 26, 2011
I am finally, and happily ditching Windows IIS, SQL Server, and ASP in favor of LAMP. Not only will I save a bunch of money on operating systems but I've found php and MySQL development to be much faster than their Microsoft counterparts.Currently I have two W2008 and two Ubuntu servers running and doing virtually parallel tasks. I want to can the W2008 machines but I am not 100% sure of my Ubuntu mirrors.Everything seems to be working fine. I've copied tons of data back and forth as a primitive test but sometimes things work fine for all the wrong reasons. Here's where I get confused.
Question 1:Do I need to partition the RAID device (MD0) and then format it?From my experience this is necessary to get the device to mount.
Question 2:In this case was it also necessary to format the individual drive partitions?
Question 3:If I do a daily cat /proc/mdstat is this all I need to do to check the drive status.
Question 4:Is there any other check I can do to assure that the mirrors are created, mounted, and operating correctly?
View 3 Replies
View Related
Mar 17, 2011
I am running Ubuntu 10.10 64. I have a RAID array consisting of two 1 TB HDD's, controlled by my on-board RAID controller. I have a dual-boot of Ubuntu 10.10 and Windows. The RAID array is mapped in /dev/mapper, and here is the output of sudo dmraid -ay
Code:
RAID set "pdc_dedfhcfdee" already active
RAID set "pdc_dedfhcfdee1" already active
RAID set "pdc_dedfhcfdee2" already active
RAID set "pdc_dedfhcfdee3" already active
[Code]....
View 2 Replies
View Related
Mar 25, 2011
I'm having trouble with Ubuntu 10.10 and stable device names. When I installed Ubuntu, the root drive was the only one in the machine; it obviously got /dev/sda.
After the base installation, I installed three additional 2TB drives to make RAID-5 array. Ubuntu renamed the root drive to /dev/sdd. While annoying I lived with it.
After creating a single partition set to "Linux raid autodetect" on each drive, I created the RAID-5 array:
Code:
All was going well until a reboot. When rebooting Ubuntu decided to make the root drive /dev/sda this time and now mdadm --detail /dev/md0 reports:
Code:
How to fix the array and make the device names stable?
View 1 Replies
View Related
Jun 29, 2011
I have a raid 0 setup with 2 x 1TB drives. I have an ASUS P8P67 LE motherboard and am using Interl RST for the Raid setup. I'm utterly ignorant of raid and therefore forgive any mistakes... I already had windows 7 installed and was attempting to dual boot ubuntu.
I installed Ubuntu from CD. The raid was picked up properly as only one drive by ubuntu. So it picked up the windows MBR and the main windows partition. I resized the main partition and used the "install ubuntu and windows 7 side by side" option. Installation went fine but once I restarted the PC I was welcomed by a grub rescue screen with the message: "error: no such device e196.....". Edit: I used the Windows 7 disc to repair the windows bootloader so I can now boot into Windows 7.
Before doing so I used gparted on the live cd to check the partitions on the drives. The only ones present were the MBR and windows one. So ubuntu seemingly didn't install... Although GRUB did... I was advised by someone on the ubuntu IRC chat to avoid trying to reinstall ubuntu at that point just in case there was an error in the partitioning process. I've since checked the state of the partitions from within windows and there's the MBR partiton, the windows partiton AND the partition that I created for ubuntu... 965MB of the partition that I created is listed as used space as well...
View 6 Replies
View Related
Mar 14, 2010
I look after some small office where computers run ubuntu. Sometimes they phone me for help. For that reason, I decided to install ubuntu alongside with my slack.I seem to have problems with lilo configuration. Ubuntu is installed on software raid :
/boot = md0 (raid 1 of sda1+sdb1)
/ = md1 (raid 0 of sda2+sdb2)
That's lilo.conf:
Code:
append=" vt.default_uft8=0"
boot = /dev/sda
bitmap = /boot/slack.bmp
[code]....
Code:
Fatal: open /vmlinuz-2.6.31-9-rt: No such file or directory
I can mount md0 without problems and the file name seems to be correct.
View 3 Replies
View Related
Feb 10, 2010
I currently have 3 drives installed to be used as a file server.
1 holds the Ubuntu OS
The other is the file server drive with 1 additional for backup using raid1.
2 Questions:
1) How do I get to the drive or Raid device to put files on the drive using command line (the 2 drives are sda & sdc that are connected to the raid1 device)?
2) How do I set the path in Samba to connect to this RAID drive.
View 1 Replies
View Related
Feb 5, 2011
I am trying to create a new mdadm RAID 5 device /dev/md0 across three disks where such an array previously existed, but whenever I do it never recovers properly and tells me that I have a faulty spare in my array. More-specific details below. I recently installed Ubuntu Server 10.10 on a new box with the intent of using it as a NAS sorta-thing. I have 3 HDDs (2 TB each) and was hoping to use most of the available disk space as a RAID5 mdadm device (which gives me a bit less than 4TB.)
I configured /dev/md0 during OS installation across three partitions on the three disks - /dev/sda5, /dev/sdb5 and /dev/sdc5, which are all identical sizes. The OS, swap partition etc. are all on /dev/sda. Everything worked fine, and I was able to format the device as ext4 and mount it. Good so far.
Then I thought I should simulate a failure before I started keeping important stuff on the RAID array - no point having RAID 5 if it doesn't provide some redundancy that I actually know how to use, right? So I unplugged one of my drives, booted up, and was able to mount the device in a degraded state; test data I had put on there was still fine. Great. My trouble began when I plugged the third drive back in and re-booted. I re-added the removed drive to /dev/md0 and recovery began; things would look something like this:
Code:
user@guybrush:~$ cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdc5[3] sdb5[1] sda5[0]
3779096448 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]
[Code]...
View 9 Replies
View Related
Mar 9, 2010
After booting, my RAID1 device (/dev/md_d0 *) sometimes goes in some funny state and I cannot mount it.
* Originally I created /dev/md0 but it has somehow changed itself into /dev/md_d0.
# mount /opt
mount: wrong fs type, bad option, bad superblock on /dev/md_d0,
[code]...
In /proc/partitions the last entry is md_d0 at least now, after reboot, when the device happens to be active again. (I'm not sure if it would be the same when it's inactive.)
Resolution: as Jimmy Hedman suggested, I took the output of mdadm --examine --scan:
ARRAY /dev/md0 level=raid1 num-devices=2 UUID=de8fbd92[...]
and added it in /etc/mdadm/mdadm.conf, which seems to have fixed the main problem. After changing /etc/fstab to use /dev/md0 again (instead of /dev/md_d0), the RAID device also gets automatically mounted!
View 5 Replies
View Related
Mar 24, 2009
configure my raid devices so it can be started and mounted at start-up.
View 1 Replies
View Related
Jun 29, 2010
For SATA drives in AHCI mode, the names are /dev/sdX, what about a RAID-0 or RAID-1 array of SAS drives? Are they the same?
View 8 Replies
View Related
Mar 28, 2011
I have 2 identical disks, /dev/sda and /dev/sdb. I have a raid-2 configuration (/dev/md0) on /dev/sda3 and /dev/sdb3 partitions. On top of /dev/md0, I am running LVM. There are also partitions sda4 and sdb4 following sda3 and sdb3 respectively but the data in there are not important. What I want to do is delete the sda4 and sdb4 partitions and extend sda3 and sdb3 to the end of disk, and grow the md0 and the volume group of course *without* loss of data.
View 3 Replies
View Related
Jan 11, 2010
I have a desktop server set up with 3 x 1Tb disks set up in a number of partitions. Some are under RAID 1, others under RAID 5. eg:
Code:
md0 : active raid1 sdc2[2](S) sdb2[1] sda2[0]
513984 blocks [2/2] [UU]
[code]....
View 6 Replies
View Related
Aug 3, 2010
We ran out of space on our server hard drive, so I installed 2 x 1GB drives, set them up as a software RAID1 array, copied the contents of /home to it, mounted it as /home for testing. Everything OK, so I unmounted it,deleted the contents of the /home folders (don't worry, we're backed up), then remounted the array. Everything was fine until we rebooted.Now I can't access the array at all; during booting the error "mount: special device /dev/md1 does not exist" comes up twice, and manually trying toe same issue.The relevant line from fstab reads:/dev/md1 /home ext3 defaults 0 0.However, using webmin shows only md0, the RAID0 device on which the OSD was originally installed.There is no /dev/md1 device file.The mdadm.conf file reads as follows:
# mdadm.conf written out by anaconda
DEVICE partitions
MAILADDR root
ARRAY /dev/md0 level=raid0 num-devices=2 uuid=76fd4050:fb820568:c9bd3a59:ad3e70b0
View 1 Replies
View Related
Jan 16, 2010
I am trying to set up a mdadm raid in a new PC that I am building for home theatre. the machine boot just fine from /dev/sdc running ubuntu 9.10 However in gparted /dev/sda and dev/sdb show to be part of /devmapper/sil_ajbicfacbaej Both dev/sda and /dev/sdb were drives that used to be part of a sil hardware raid on a previous machine. I would like to use them as a new mdadm raid on this new machine the old hardware card was really quite slow. the drives are now pluged into the MB and should bw much faster there.
fdisk -l shows this
*********************************************** ~$ sudo fdisk -l
Disk /dev/sda: 1500.3 GB, 1500301910016 bytes
255 heads, 63 sectors/track, 182401 cylinders
[code]....
View 6 Replies
View Related
Dec 16, 2009
Got a little problem after the install of Fedora 12. First there was not problem in opening the raid-device, after i tryed to automount it with crypttab and fstab im not longer able to open it.
Here some outputs code...
View 2 Replies
View Related
Aug 14, 2010
I am hoping there is somebody out there who might recall DPT hardware and be able to get it to work under Linux. DPT were bought up by Adaptec but made great SCSI cards. The PM3334UW is a true hardware raid device which I have been using in SOHO environment for the last 7 years or so with IBM Warp 4.52 (MCP2). The box, although legacy hardware is sound and I wish to keep it running but hoped to have an openSUSE option. Unfortunately I cannot find a driver for the SCSI RAID card. There are a good deal of posts concerning eata_dma.c and other eata based drivers as possibles but when I load the eata driver that is available when loading openSUSE 11.3 I am asked for parameters and I haven't a clue what these should be.
View 4 Replies
View Related
Mar 19, 2011
I just tried creating a raid Device which has both stripping and mirroring i have done as below Quote:
mdadm -C /dev/md00 -l1 -n2 /dev/sda7 /dev/sda8
mdadm -C /dev/md01 -l1 -n2 /dev/sda9 /dev/sda10
mdadm -C /dev/md02 -l0 -n2 /dev/md00 /dev/md01
pvcreate /dev/md02
vgcreate volgroup02 /dev/md02
lvcreate -n orac -L 9G volgroup02
[Code]...
Everything is fine until here but after reboot the device wont mount on /orac it says special device not available i found that that md02 device is not in active state
i tried deleting it and recreating it but no use still it wont persist a reboot
View 1 Replies
View Related
Dec 4, 2010
I have an Linksys NSLU2 with four usb harddrives attached to it. One is for the os, the other three are setup as a RAID5 array. Yes, I know the raid will be slow, but the server is only for storage and will realistically get accessed once or twice a week at the most. I want the drives to spin down but mdadm is doing something in the background to access them. An lsof on the raid device returns nothing at all. The drive are blinking non-stop and never spin down until I stop the raid. Then they all spin down nicely after the appropriate time.
They are Western Digital My Book Essentials and will spin down by themselves if there is no access. What can I shutdown in mdadm to get it to stop continually accessing the drives? Is it the sync mechanism in the software raid that is doing this? I tried setting the monitor to --scan -1 to get to check the device just once, but to no avail. I even went back and formatted the raid with ext2 thinking maybe the journaling had something to do with it. There are no files on the raid device, it's empty.
View 12 Replies
View Related
Aug 20, 2009
In a nutshell, our RAID 1 array was rendered broken and we were advised that core lib files were missing and the OS needed to be reloaded... a quote from our server host:"The OS is not healthy.This server will need a reinstall.
Libs are missing." This was after having replaced what we though was a faulty /dev/sdb. So they reloaded the OS (Debian 5.0.2 x86_64) on 2 FRESH drives, and installed the old /dev/sda as /dev/sdc once the reload was completed. Here's the output of /etc/fstab on the fresh install so we know what we're working with:
Code:
debian:/BAK# cat /etc/fstab
# /etc/fstab: static file system information.
#
# <file system> <mount point> <type> <options> <dump> <pass>
[code]....
The one problem I see myself running into is /dev/md1 and /dev/md2 are currently in use by the new system, so I cannot mount it there. I should also note, reloading the OS is a viable option if needed as we haven't started configuring the server yet. So if we need to reinstall the OS and assign the NEW RAID arrays to something other than /dev/md1 and /dev/md2 then we can do that.
View 3 Replies
View Related
Sep 9, 2009
Alternate Heading: Unable to use opensuse partitioner to fully format a Seagate 1Tb drive OK. Swapped over a motherboard to increase the number of slots available for hard drives so that I could expand my raid array (4 X 1Tb drives). Discovered I had no thermal paste so all delayed for 24 hours while I bought some, miscalculated on rebuild and had to reinstall OpenSuse but in the end system is now up and running. Unfortunately when I formatted my new 1Tb drive (Seagate) it formats to 931.50 Gb while the other 4 drives formatted to 931.51 Gb (they are WD). I'm now in the position that when I try to add this new drive I get :
mdadm: /dev/sdc1 not large enough to join array
Is there any way I can resize the existing devices down to 931.50 Gb so that I can add in the new drive without having to restore the array?
View 4 Replies
View Related
Oct 30, 2009
I'm writing a bash script that needs to know whether or not a device node is part of a RAID array. I'm just curious if anyone knows of a good way to determine if a device node is in a RAID array. I know that you can run mdadm -Q or mdadm --examine on the device node and that will tell you. But I don't want to rely on screen scrapping and would rather have something that would return a boolean. Any ideas?
View 2 Replies
View Related
Sep 9, 2010
Recently did an apt update and upgrade to my CLI only Lenny server. Upon reboot I get an "ATA softreset failed (device not ready)" for all of my SATA drives. I noticed the upgrade changed the kernel to "Linux debian 2.6.26-2-amd64" (do have 64bit CPU).Once loaded to a command prompt I can assemble my raid 6 array with the command "mdadm --assemble /dev/sda to sdd" then mount it with mount -a. But transfers to the array areorribly slow ~1mbs.Upon reboot i get the same errors and have to assemble my array every time
View 1 Replies
View Related
Aug 3, 2010
We ran out of space on our server hard drive, so I installed 2 x 1GB drives, set them up as a software RAID1 array, copied the contents of /home to it, mounted it as /home for testing. Everything OK, so I unmounted it, deleted the contents of the /home folders (don't worry, we're backed up), then remounted the array. Everything was fine until we rebooted. Now I can't access the array at all; during booting the error "mount: special device /dev/md1 does not exist" comes up twice, and manually trying toe same issue. The relevant line from fstab reads:
/dev/md1 /home ext3 defaults 0 0
However, using webmin shows only md0, the RAID0 device on which the OSD was originally installed. There is no /dev/md1 device file. The mdadm.conf file reads as follows:
# mdadm.conf written out by anaconda
DEVICE partitions
MAILADDR root
ARRAY /dev/md0 level=raid0 num-devices=2 uuid=76fd4050:fb820568:c9bd3a59:ad3e70b0
So it's not listed; I'm assuming this is significant. Am I right, and whether I am or not, what can I do?
View 3 Replies
View Related