Fedora Hardware :: Software RAID1 /boot Volume Doesn't Mount Automatically At System Startup
Feb 7, 2010
My software RAID setup is as follows.
/dev/md0 (made from sda1 and sdb1) RAID1 /boot partition
/dev/md1 (made from sda2, sdb2, and sdc2) RAID5 / partition
Earlier on I had some trouble with my sda drive, it dropped itself from both arrays, screwing up the mirroring of my two raid partitions participating in the /boot partition. I eventually got everything sorted out and back in sync. (I also have grub installed to MBR on both sda and sdb). Things are working fine regarding that, but since then I've had this issue:
During boot up, I'll get an error message that it could not mount my /boot partition (when fstab is set to either /dev/md0 or the UUID). It claims c9ab814c-47ea-492d-a3be-1eaa88d53477 does not exist!
[mark@mark-box ~]$ cat /etc/fstab
# Created by anaconda on Wed Jan 20 16:34:41 2010
As far as I know, it isn't neccessary for /boot to be mounted always, correct? Although, as I understand, I need to have it mounted whenever making kernel changes correct?
I'm having a strange behaviour with GRUB, and I don't know much about it. The problem is that GRUB is looking at the wrong place when it starts. I have a /boot partition (/dev/sda1) with everything for booting and grub. This gets updated when I update, say, the kernel, grub, etc. But I also have a /backups partition (/dev/sda4) with a directory /backups/boot/grub that is a copy of /boot/grub. The /backups/boot directory only has the /backups/boot/grub directory, not the kernels, etc.
If I delete the /backups/boot/grub directory, I get a "grub>" console at startup, and system doesn't boot. SO: Why is grub looking at /backups/boot/grub instead of /boot/grub and how can I fix this? This is my grub.conf, identical in both directories, and both pointing to /boot=/dev/sda1=(hd0,0):
One question: should F13 mount all attached USB devices after boot automatically? I guess it should. However, what I've experienced is that after boot and login, my USB modem + flash memory is not mounted. I need to manually unplug it and plug it again, and then it's mounted
I have an mdadm/lvm2 volume with 4 HDs that I created in Ubuntu 10.04. I just upgraded the computer to Ubuntu 10.10.
I redid the mdadm commands to get volume up and running, did mdadm --detail --scan > /etc/mdadm/mdadm.conf to get the configuration file.
But now, every time I reboot, it tells me that the volume is not ready. /proc/mdstat says that I always have one disk of the volume "inactive" as md_d127. I need to stop this volume and reassemble the whole thing to get it working.
This is what I get out of mdadm --detail --scan and put inside /etc/mdadm/mdadm.conf:
And this is my /proc/mdstat on boot:
I need to do mdadm -S /dev/md_d127, mdadm -S /dev/md127, mdadm -A --scan to get this volume working again.
This did not happen with Ubuntu 10.04. I'm really fearing the loss of my raid5 data now.
My pc has two operating system, XP and Fedora 14 I was setted PC to Auto turn on with BIOS, after turning on, system was booting in fedora automatically after 10 seconds.
But, After i upgrade to fedora 15 via terminal(Preupgrade), that auto booting has disabled. I should Press Enter to boot into Fedora 15. It should be have auto booting into feodra 15 like when i was fedora 14.
I had to re-install Windows XP because the install was running slow.So, I created another partition using GParted for my personal data and moved my files there and re-installed Windows XP.Now, the Windows partition won't mount automatically.NTFS Configuration Tool shows 0.0GB. So, I have to open up a Terminal window, and issue sudo mount /dev/sda1 /media/Windows and everything is fine.
I It does not start when the system boots. The smbd daemon is present and there are no suspicious messages in the system log, dmesg or the samba logs. The nmb service *is* enabled at all of the relevant run levels and does start if I issue the command "service nmb start" as root after I log in. This particular box's networking is 802.11n via a usb port and is under the control of the network manager. I have another fedora 12 box with an internal intel 5100 802.11n card and it does not have this problem.
I have just upgraded to Fedora 14 from an older version. I now have problems mounting my RAID1 array, which was operating correctly until now. This is a software RAID which was initially built under Fedora 10.The array is md0, and is made of 2 SATA drives (sdc and sdd) which have only one partition. The underlying filesystem is NTFS. The array is assembled correctly and active, as reported by /proc/mdstat and mdadm -D.When I try to mount the array, I get this:
Code: [root@Goofy ~]# mount /dev/md0 /mnt/raid mount: you must specify the filesystem type
I am struggling with getting an sshfs mount mounted on system boot. I have a script that mounts the sshfs for "userA". When userA runs the script all is well - user A can access the remote filesystem, root user can't see it as expected. The basic command is: sshfs userA@remote host:/home/userA /home/userA/mountdir -p 21212 -o password_stdin < passwordfile. I can prepend the sshfs command in the script with su - userA -c and when I run this script logged in as root all is well, userA has access and all is well. If I then put this script in /etc/init.d and reference it properly in the rc. directories the mount doesn't happen. If I prepend the sshfs command with sudo, same thing. Logged in as root I can run the script and UserA has access. Run the script in /etc/init.d during startup and the mount doesn't happen. Echoing text to a log file shows that the script is being executed but no mount happens.
How-To set the mixer volume level at system startup or login. A funny story that led up to this how-to first...
I made the obvious mistake of leaving my volume level set to nearly maximum. Of course, Ubuntu's default behavior is to restore the mixer to its last known state - a point of much irritation at that moment. This had been a problem in the past as well, and today was the last straw. So, I did some research, for quite some time I might add, and decided to be a good community member and share my findings. It seems that there are all sorts of opinions around the web. The dominant opinion is that a mixer should always be restored to its last known state, that this is all well and good, and why would you ever want it to work any other way. Lots of people suggested that the startup sound be disabled, which was not a terrible solution, but was still a work-around as it means that the next sound bite to be played is the alarming one.
Needless to say, I wanted to find out what I would call the "proper" way to set the mixer level at startup. As my laptop uses PulseAudio, and my office desktop uses ALSA audio, the methods were different. My focus was for PulseAudio as that was the original purpose. I note here that my method for ALSA is less detailed as it is not the default for Ubuntu audio these days. So, if you are using ALSA, you might have to be a little creative to make my ALSA note fit your needs. I have attached 2 files to this post, one for ALSA and one for PulseAudio.
The other day a power outage affected an SSH server I have running 9.04 (32-bit desktop ed.). I have two external USB hard drives used for cron-scheduled backups (one for rsync, one for an incremental) that are connected at all times. When I reboot, they no longer mount automatically until I login to the gnome desktop. As far as I can remember, they always mounted automatically before as disk-1 and disk-2, but now I have to login to the gnome desktop and then log back out.
I never had them listed in the fstab before since they just worked, & hope to avoid doing it this way since the drives sometimes get their paths (sdd and sde) interchanged. However, is the best way to fix this to use UUIDs in fstab vs. using sde or sdd? (such as in this post: 4highlight=external+hard+drives+don't+mount+boot Or maybeust remembering incorrectly & ubuntu doesn't mount them automatically until login? Sometimes I have to reboot remotely, & this problem would cause the rsync to fill up my system drive.
Is there any way to automatically activate the wifi at system startup without using the keyboard switch? I have tried this script in gedit, saved as .run and added to startup applications but it doesn't work:
my "places menu" references folders on a second partion, where I store all my personal files (ntfs, as I use it also from Windows). After boot those folders wouldn't show in Thunar. Some other programs, as Fontmatrix, wouldn't find databases stored there, what messes them up. Thunderbird has problems to recognize profiles etc.What is strange, the partition seems to be mounted at startup as I can see and select it from Thunar. Also, it appears on Desktop. After selecting the partition, closing and reopening Thunar everything works fine, places would show correctly and Fontmatrix database is ok
I have installed Oracle 11g on 5.3. Now, I am trying to START and STOP oracle automatically on system startup and shutdown. I am following this link To have this effect, I have created /etc/init.d/dbora
I need a FREE solution that can image an entire Luks system encrypted volume and the rest of the used HDD, the MBR and /boot partition. Note: MBR and /boot are not encrypted. Note 2: I want to be able to restore entire drive from image with only a couple of steps. Note 3: Destination HDD space is a factor. Image file must be compressed and the image file must be around 40 to 50 GB or less. The smaller the image the better.
I have used clonezilla live cd before but not for encrypted volumes. I know you can install it in Linux. But, I don't know how to configure it after installation. I would be very happy if someone could tell me how to configure clonezilla in Fedora. How to guides are also welcome. I have one more question. If I image the encrypted volumes and all the stuff I mentioned above while logged in to Fedora, and I restore the drive from the image, will the recovered drive still be encrypted?
I have trawled through an extensive number of post on quite a few forums without even a step forward with this.
I have a fedora 13_x64 system with software raid1 for /boot and / (md0 and md1 respectively) , swap is not raided.
I was doing an yum update through the software updater in gnome and the system froze.
I had to press reset to get any response from the machine.
Since then I have been getting the kernel panic above just after grub starts fedora.
I tried the previous kernel from the previous update and it has the same error.
At the worst I am prepared to load OS again but there is still some info and configs that I would like to access from the raid partitions before I go ahead. Is there any way to access these partitions through a live CD or rescue environment?
Is there a method to bring this install back to life? or am I looking at a reinstall?
I believe server section is the best when speaking of RAID stuff...
I have the following situation:We have a DELL T3400 with embedded fake raid on it. I dont know exactly how the system was setup (I wasnt here at that time), but the RAID was enabled in bios and while booting, the two harddrives would be seen as members of intel raid volume0 (RAID 1 mirror). I am not sure if the software raid was actually properly configured in Linux (Fedora 9) and if the OS was reconstructing the whole raid or it was just the bios part that was mirroring the /boot or just some parts of it. Frankly I find these hydrid raids very confusing.Some bad disk manipulation from my part caused the server to crash, but I was able to recover and boot just with one hard drive after using fsck.
I decided to get rid of the raid as it's not the right solution for the application we need it for and decided to go for a traditional single harddrive system and to use Ghost for Linux to clone to a spare disk when backups are needed.So I installed the latest Fedora 12 distribution onto another harddrive and disabled RAID in bios (changed from RAID ON to autodetect, which is the only other option).
Here is what I have now: /dev/sda has the newly installed fedora 12 /dev/sdb is an empty harddrive that I would use as an intermediate /dev/sdc is the old harddrive member of intel raid volume0
sdb was partitioned into sdb1 sdb2 and sdb3 and I created an ext3 filesystem on sdb2. The hard drive belonging to RAID volume0 (sdc) has a lot of work done on it and I would like to be able to recuperate the files to the new disk (sda). I cannot mount that old harddrive while in fedora 12, as it sees some unknown raid member filesystem on it probably assigned by the intel raid chip.
So I decided to do it from the other side: to boot from raid volume 0, and from there mount a third intermediate harddrive (sdb) onto which I would copy the documents and then mount the same harddrive from the newly installed fedora 12 and copy those documents from that intermediate harddrive.I can mount /dev/sdb2 from fedora 12 fine and copy stuff to and from it, but not when I boot from the RAID volume 0 harddrive (sdc) with fedora 9 on it. It keeps saying that the partition in question (/dev/sdb2) is an invalid block device.I am stuck here, as my knowledge in this sort of things is very limited.If somebody can indicate me how to recuperate files from that old raid harddrive onto the new fedora 12 drive, I would appreciate a lot.
I have an existing Fedora 15 system installed from scratch.I've ordered a harddrive identical to my SDA and want to add it to my existing system as a RAID1 setup.I've googled around and cannot find recent clear instructions how to accomplish this. I don't want to reinstall everything from scratch. It should be possible to create the RAID1 using the existing data disk and then mirror everything up?
I have just installed 10.04 and am enjoying the quick clean boot. I have a dual-boot option as I have come over from XP. Most of my data is still on the old 'c:/' drive and I have to manually mount it every boot-up. Is there any way to automate this process?
I am trying to boot my F13 server that has 3 partitions (sda2,sdb2,sdc2) configured as RAID1 (md0), vg00 is on md0 and / is on vg00/lvol00 and boot is on /dev/sda1, sda3,sdb3 and sdc3 contain other non-OS data and I get the following errors dracut scanning devices sda3,sdb3,sdc3 for LVM logical volumes vg00/lvol00 vg00/lvol01 . dracut: Volume group "vg00" not found dracut: Skipping volume group vg00 dracut: Autoassembling MD Raid No root device round and then everything stops If I go in with rdshell and type in -
Ok I managed to solve it. When I originally installed F13 I had 2 partitions in md0 and later on I added /dev/sdc2 to make it a 3 partition Raid 1 array. My thinking was seeing that I had the extra space it wouldn't hurt to use it as another mirror. One month later after I booted the server I have found out it did hurt. I removed the 3rd partition and now it's fine again. I wonder though if it would be possible though to use that 3rd partition.
How to mount vfat partition automatically after boot? After login it it will mount all vfat partition and the icon of those parition will be at desktop. How can it be done. udisks is installed. If i click a vfat partition from pcmanfm it prompts for password to mount.I don't want to click. It will be automatically mounted and i will get the icon of that mounted vfat partition at desktop
I recently switched from fedora 14 to 15. Today my computer suddenly shut down during an update, as I thought it overheated I decided to clean the cooling system and reapply thermal paste. However, now the system won't boot anymore ("kernel panic - not syncing : VFS: unable to mount root FS on unknown-block").
I would like to either solve this booting problem, or mount the fedora 15 filesystem and recover some files. Whichever is easier.
I have another drive with fedora 14 (antec below) which boots fine:
I have recently installed denyhosts to help guard against bruteforce ssh attacks on my Fedora 12 server from the Fedora repositories. If I manually start denyhosts (as root) using: /usr/sbin/denyhosts.py --daemon
it works fine. The denyhosts log file is created and indeed the /etc/hosts.deny file is updated. However if I make attempts to start denyhosts automatically upon the server restart denyhost fails with a permission denied error for /etc/hosts.deny (error 13) I have tried using chkconfig to enable /etc/init.d/denyhosts on run levels 3, 4 and 5. I have also tried including the line:
in /etc/rc.local Both of these attepts report the same error. anyway to automatically start denyhosts or know why this problem might be happening?
i'm trying to get everything working ok. i have installed ubuntu using wubi and i've found that i can access my files on my windows partition from ubuntu. to do this i have to mount the disk and enter the password each time i boot up, and i would like this to be done automatically. i was wondering if this was possible? i put in a link directly to the music folder on windows into my 'places' but it only appears once i have put the password in. its not a huge thing, but its one of those things which would make starting up my ubuntu a lot more conveniant.