Red Hat / Fedora :: Extending Root Partition - Volume Group Not Found
Sep 20, 2010
I am trying to extend my / size as its full. Well the volume group is VolGroup00 & logical volume is LogVol00 but when. I run the command vgextend VolGroup00 /dev/sda8. It says volume group not found. Can it be because I have WindowsXP in my /dev/sda1, which falls under same Volgroup??
Dual PII 400, 512Mb with a Promise SuperTrak 100 IDE Array Controller. At present I have only one drive on the controller, configured for 1 JBOD array. I install FC9 with no problem. New partition is created and formatted, Grub is installed, and then... Grub is found and booted, but then I get:
Reading all physical volumes. This may take a while... No volume groups found Volume group "VolGroup00" not found Unable to access resume device (/dev/VolGroup00/LogVol01) mount: could not find filesystem '/dev/root' I can boot in rescue mode, chroot to the installed system. I changed the kernel boot parm "root=/dev/VolGroup00/LogVol00"
I have serwer Debian with my website. My provider splited the disc into 5GB partition for / and 495GB partition for /var. Everything was going ok for over two years but now I don't have enough memory on /. I'd like to increase the partition but the problem is that /var is just next to it so I can't easily change the end of the first one. I need some safe solution. It might be even just shrinking partition for /var, adding new one after if it helps anyhow (I have about 450GB free memory).
# parted print GNU Parted 2.3 Using /dev/sda Welcome to GNU Parted! Type 'help' to view a list of commands. (parted) print Model: ATA ST3500418AS (scsi) Disk /dev/sda: 500GB Sector size (logical/physical): 512B/512B Partition Table: msdos
Number Start End Size Type File system Flags 1 512B 5369MB 5369MB primary ext3 boot 2 5369MB 500GB 494GB primary ext3 3 500GB 500GB 538MB primary linux-swap(v1)
Ok so I have one drive. /boot /lv_root and /lv_swap
At the end of the drive I have 32 gigs of free space still contained in the logical volume group. I want to remove it from the LVG but this is on one device. Supposedly there is a way to do this, pvresize and fdisk.
[URL]
Quote:
Originally Posted by source
#I've tried to shrink the PV with pvresize which didn't throw errors -
Good.
#but fdisk still shows me the same LVM partition size as before.
That's normal. pvresize "just" updates the PV header and VG metadata.
#So I guess the partition table has to be modified somehow?
Yes. That was mentioned in my reply: "Then shrink the partition in the partition table."
You can use fdisk or any other partition table editor for this. Some don't support resizing a partition. In that case, you can delete and create a smaller one. If doing the delete/create dance, you *must* create the new partition on the same cylinder boundary as the current one to preserve the current data.
Ive read from every source on LVM its not possible to do this. Why on earth would any Linux developer put LVM on a single drive system by default? Were they even paying attention? I dont mean to go off on a rant but if there are multiple drives LVM makes sense. However if you only have one large drive LVM holds your system hostage and you have to crawl thru the pit of hell to get it back.
I understand you have a choice in the matter when you install Fedora but its really the worst possible choice for default. Many newcomers to Linux run into this problem with LVM. If you cannot resize LVG's the software should have never been put into a Linux distro in the first place.
I have a dual boot system running FC8. I had a problem with an uninstaller on the Windows side that affected both the drives on the Windows and Linux sides. It froze my Linux archive drive. I removed the drive, and was able to boot the Windows side from a recovery disk, but not through Grub. I can't access Grub, but can start the Linux boot process with SuperGrub. The boot process gives the following errors:
No Volume groups found. Volume group "VolGroup00" not found. mount: could not find file system 'dev/root'.
I have run into some serious problems trying to start up RHEL AS4. I am trying to install Oracle on this box which by the way is running as a guest OS through VirtualBox. I was running with it fine yesterday. I was doing the prerequisites of updating the /etc/sysctl.conf file with additional kernel parameters. Prior to that I added another scsi virtual disk and extended a PV and left it at that for the night. I am since come back today and trying booting and am getting the following errors:
Volume group "VolGroup00" not found; Couldn't find all physical volumes for volume group VolGroup00; Couldn't find device with UUID "Some UUID of the device"; ERROR: /bin/lvm exited abnormally! (pid 318); mount: error 6 mounting ext3; mount: error 2 mounting none; switchroot: mount failed: 22 umount /initrd/dev failed: Z kernel panic - not syncing: Attempted to kill init!
I have tried running in rescue mode and it can find volume group VolGroup00. I also have installed Storage Foundation which includes veritas volume manager and other various veritas components. Does anyone have a clue what I am supposed to do here to get RHEL up and running again? I am running kernel version 2.6.9-42.EL if that helps as well.
Before creating this topic I googled a lot and found lots of forum topics and blog posts with similar problem. But that did not help me to fix it. So, I decided to describe it here. I have a virtual machine with CentOS 5.5 and it was working like a charm. But then I turned it off to make a backup copy of this virtual machine and after that it has a boot problem. If I just turn it on, it shows the following error message:
Activating logical volumes Volume group "VolGroup00" not found Trying to resume from /dev/VolGroup00/LogVol01 Unable to access resume device (/dev/VolGroup00/LogVol01) ... Kernel panic ...! During the reboot I can see 3 kernels and if I select the 2nd one the virtual machine starts fine, it founds the volume group etc. (But there is also a problem - it can not connect the network adapters.) So, it is not possible to boot it with the newest kernel (2.6.18-194.17.1.el5), but it is possible with an older one (2.6.18-194.11...)
I looked into GRUB's menu.lst and it seems to be fine. I also tried #mkinitrd /boot/initrd-2.6.18-92.el5.img 2.6.18-92.el5 no luck! Yes, I can insert DVD .iso and boot from it in "linux rescue" mode.
This is a problem about linux-kernel-3.16-0-4-amd64 and LVM, I guess. I decided to write this here in case other users who installed their debian system with encryption enabled experience this problem with a recent kernel upgrade.
I use debian jessie. Today I gave the command:
Code: Select allapt-get upgrade
There was a linux-kernel upgrade to 3.16-0-4-amd64 among other packages to be upgraded.
After this upgrade my computer cannot boot anymore.
I get following error:
Code: Select allVolume group "ert-debian-vg" not found. Skipping volume group "ert-debian-vg" Unable to find LVM "volume ert-debian-vg/root" Volume group "ert-debian-vg" not found Skipping volume group "ert-debian-vg" Unable to find LVM "volume ert-debian-vg/swap_1" Please unlock disk sd3_crypt:
And it does not accept my password.
I used rescue environment on debian jessie netinst iso and decrypted the partition and took a backup of my /home. Now I have not much to lose if I reinstall my system but I still want to fix this problem if possible.
I have reinstalled the kernel using debian jessie netinst rescue iso but nothing changed.
I have Timeshift snapshots located at /home/Timeshift but timeshift --rescue command cannot find a backup device, it sees the device as crypted. If I could restore a snapshot it would be very easy to go back in time and get rid of this problem. It would not be a real solution, however.
There is not any old kernel option in GRUB menu. So removing the latest one does not seem as an option.
I don't know much about lvm and I've managed to screw up a drive. I had a 500GB drive with FC14 on it and I wanted to copy over a load of data to my new 1TB that was replacing it. I set up my new install the same way as the old...including the same volume names (error number 1 I think) I successfully mounted the old/500GB drive (using vgscan and vgchange -a y etc.) using a laptop (running FC13) and an external hdd cradle. I could access the files I wanted but this wasn't the machine I wanted to copy them to (I was doing this while waiting for the install to finish on the new drive).
When I tried the same process on the new install I found that having two lvm with the same name meant I couldn't mount the external one. So I opened the disk utility (palimsest) and was going to change the name of the old volume group but it wouldn't let me do that. I then thought maybe I could get away with just changing the name of the partition where the files were and maybe I could add it to the mounted group or something so I changed it to lv_home2. This changed the name of my new/1TB lv_home to lv_home2 as well. So thinking that wasn't the answer I just changed the name of the new lv_home2 back to lv_home.
From that point on I haven't been able to see the old drives partitions (the new volume group still works so far). I has a physical volume but the volume group and volume names are gone from view. When I try to vgscan on my main computer or the laptop I had it working on earlier I get:
I have installed oracle enterprise linux on VM ware with 20 gb allocated to guest OS. Now I want to install oracle apps in the guest Os, so I need to extend the volume. I have extended in Vm , but I have to partition in the guest OS, for that purpose I am using Gparted. But I am unable to extend to sda1. I need to have all the unallocation space allocated to sda1. Here is the screen shot, how can I do that. Right now when in press the command df -h in terminal I am gettig 18 gb as space available for sda1, I want to make it 200 gb, in which I would like to install oracle apps. Check out my screen shot.
I have taken over looking after a EMC/Suse iscsi storage network. We recently run out of space on our virtual disk, so I added another 100GB from our storage pool using the EMC management software and the virtual disk the suse box is using expanded from 1.5TB to 1.6TB. That completed fine.
However the SuSe host is still showing 1.5TB. Do I need to extend the volume somehow? Or is it just a case of restarting the suse box. I have umounted and re-mounted but no change. It is imperitive the current data on there REMAINS INTACT!
This is my development virtual machine, I had Centos 5.3 on it and a lot of stuff I really can't afford to lose.
First off I'd like to say if there is any method anyone knows of recovering information off a VMWare disk, I'd love to trash it. I've already tried the drive mapping in VMWare workstation but only the boot partition comes up and I need /home and /etc.
I was screwing around with the LVM size of the disk and totally screwed it up. I can see the two partitions on /dev/sda, /dev/sda1 as the small partition and /dev/sda2 as my large data partition (the one I need access to).
If anyone could guide me into getting my data off or rebuilding my LVM to get it booted again that'd be amazing.
I already know "what" my problem is, however I am having difficulty fixing it. I recently upgraded our companies server to a HP ML150; decided to upgrade to FC10 hoping it would go smooth and it is not. It does not detect the SATA drives after the installation.
I get.
"Volume Group "VOLGroup00" not found Unable to access resume device (/dev/VolGroup00/LogVol01 mount: error mounting /dev/root on /sysroot as ext3: No such file or directory
I know the problem is that my SATA is not enabled in the kernel or grub, but I don't know how to fix this. My internet searches are coming up a little short and LIVE discs are not working so I am having trouble figuring this out.
I have a question about LVM. My /dev/sda disk is partitioned into Windows NTFS on sda1, Linux /boot partition on sda2, and the Fedora 10 root (/) LVM partition is on sda3. I have moved my Windows XP to VMware on the Linux system and would like to add the sda1 partition to root LVM group.
A bit of an oddity that I've recently run into with my storage folder in my system; it's a newly installed drive that I've set to mount at /storage. When I first tried to use it, programs that I used that attempted to write to it tossed Access Denied errors at me in their own way. Checking the permissions (at the Terminal, ls -l / | grep storage) showed that /storage was set to 'rwxrwxr--'--Owner and Group were given full read/write/execute, but Others could only read. However, my logon to my system is a member of group root. Why, then, with the above bits set, would I not be able to write to it? Changing Others permissions to rwx (and presumably rw would have worked out for me since I don't leave anything executable there) allowed me to write to it, but I don't understand why that would have been necessary. So far as I'm aware, the prior drive that was in my system--mounted at the same location--did not need this treatment.
I have one HD and a VG spanning it's entirety. I resized a LV and freed up 10GB within the VG, but I want the 10GB outside the boundary of the VG so I can install another OS for testing purposes. For some reason I'm not able to do this. I don't know if I understand LVMs correctly. Maybe there can be only one VG on a HD?
I'm rearranging a bunch of disks on my server at home and I find myself in the position of wanting to move a bunch of LVM logical volumes to another volume group. Is there a simple way to do this? I saw mention of a cplv command but this seems to be either old or not something that was ever available for linux.
1) Why would I create a new volume group to add a new hard drive to a system, rather than add the drive to an existing volume group?
2) If I created a new volume group and added a new hard drive to it, would I see the total free space (I see 30 GB now via the file browser)? For example, if I have 30 GB free on the main drive (with the OS), and I add a new drive of say 40 GB in a new volume group (using LVM) would I see 70 GB of free space? That doesn't seem to happen.
Fedora 11. I am trying to setup kickstart so it lays out a mirrored volume group. I have 2 disks sda and sdb. I want a primary partition on each disk 200mb in size for /boot. This is to be mirrored onto raid device md0 (raid 1). The rest of each disk is to be setup partition which grows to use the remaining space, and is also mirrored (raid 1) md1. Onto md1, I want an LVM volume group called rootvg, and logical volumes set up on there for /, /home, /usr, /tmp etc. I can lay this out manually, and it works fine. However, the code below, which is slightly amended from a previous anaconda-ks.cfg file doesn't work.
Code: clearpart --linux --drives=sda,sdb --initlabel part raid.11 --asprimary --size=200 --ondisk=sda part raid.12 --grow --size=1 --ondisk=sda part raid.21 --asprimary --size=200 --ondisk=sdb part raid.22 --grow --size=1 --ondisk=sdb raid /boot --fstype=ext3 --level=RAID1 --device=md0 raid.11 raid.21 raid pv.1 --level=1 --device=md1 raid.12 raid.22
i would like to extend my main file system into the unallocated space that i have on my hard drive, the unallocated space is most of it, as it used to be a partition but was deleted, do i have to do this with a boot up disk because i think that it can only be done on an unmounted partition, or is there a way to do this while linux is running in the main partition.
Is it possible if I am only using ext3 and no LVM or anything else to re-size the partition into another physical device? I am pretty sure the answer to this is no but I was still curious as I am facing a full 1tb disk and need to add a new drive and unsure how to do this due to shared folders existing on the old drive and no way to actually expand them without linking in new files or something.
I'm still pretty new to servers and ubuntu and have ran into something I could see being a problem in the future. On a dedicated remote server I have installed a web server using the "How To Forge - Perfect server set up for ubuntu 10.04 and ispconfig". I have a forum and email up and running and shoutcast radio and teamspeak3 servers also. We can also nx into it if need be. I can reformat the 1T hdd remotely from my provider control panel and ssh is installed at the same time and a new root password is sent via email. The thing is I now realise that by default the file system is written to a 10gig partition.
This might usually be ok but ispconfig uses the /var/www folder on the file system to house the forum I host and the partition is filling up. My mate I co rent with is talking about starting a parrot/bird owners forum and i might eventually like to have a gaming forum as well. I realise I should probably have set things up differently but like I said I am new at all this and tbh the home directories never going to have much in it so theres 900 gig doing nothing. So my question is can I use anything to enlarge that partition remotely? I know theres gparted on disc and all the articles I found say I need to use a disc which obviously is out of the question.
So what I think I need is some sort of partition magic for ubuntu. I would really like to expand it so all my current files etc on it would stay as is. I'm also currently looking into back up methods and wondered if that would be the way to go? Back up my file system and home directories and then reformat and make the partition larger? Or could I copy the entire www folder to a newly created folder in /home and re write the site enabled files to point to it? Would this work and if so what else would I need to edit?
I installed fedora 13 64 bit and it works great but I encountered several issues when setting up guest OS with KVM. The problem seems to be related to selinux. But let me first ask question about logical volume. By Default fedora created logical volumes:
[Code].....
"If you expect that you or other users will store data on the system, create a separate partition for the /home directory within a volume group. With a separate /home partition, you may upgrade or reinstall Fedora without erasing user data files." seems to suggest I have to create a separate physical partition and assign that to /home. But reading elsewhere it seems to suggest logical volume acts like a partition. My goal is to make it easy in case fedora is hosed and I have to re-install it without affecting /home where my cirtical data resides. Given above do I need to create a separate physical partition or I am just fine?
I have a second hard disk that originally had windows and all my data. Windows is hosed but I can see my data from within Fedora and Windows is gone and I created created new partition in its place which used ot be the C:/ drive appears as 53 Gb filesystem. My data which was originally D drive appears as 215 GB filesystem. As given in [URL] I want to create a new logical volume in 53 Gb filesystem which I want to use as space for virtual disk to install guest OS's in KVM. Currrently 53 GB filesystem is mounted as /media/3467BH89JK789 but this does not work well with KVM. how do I create this logical volume out of 53 Gb filesystem partition and add proper selinux info and do I add to vg_vostrolx volume group and in a different volume group?
The normal user is now in the sudoers group. How can i allow it to install programs using it's own password rather than having to know the super-secret Root-Users password?
My computer: (Lenovo T61 Thinkpad, running fc11 for about 2 and half months). Apparently I when I made my partitions I didn't leave quite enough room in my root directory, because I just completely ran out. Here is how my hard drive is partitioned:
The root had about 15 gigs on it, which just filled up. When I restarted to see if that would help, when it rebooted it went fine up to the log-in screen. Instead of the usual fedora blue background, it was black except for the log-in window, which looked very low-res. A little pop-up kept coming up saying the GNOME power configuration settings failed to load or something. When I logged in, the whole screen was black except for the mouse, and I could get no response. I have plenty of space left in home, so I rebooted to rescue mode using the first fedora installation disk, and tried the following command:
Code:
lvreduce -L90G /dev/mapper/DRIVE
which only returned:
Code:
lvreduce: relocation error: lvreduce: symbol dm_tree_node_size_changed, version Base not defined in file libdevmapper.so.1.02 So I couldn't reduce the size of home, and thus couldn't increase the size of root.
IN SUMMARY:
a) the lack of memory in root the probable cause for my computer not working
b) there a good way to reduce home and increase root while running this live disk
Note: When I am looking at it now in the logical volume manager, it says that on the whole physical volume there is only 400MB free. However, when I last looked (about 30 mins before I started having problems) it said there were about 100 Gb free.
Edit: Nevermind. I did some more research and it turned out to be more of a gnome power manager thing rather than a memory space thing, although I'm certainly going to increase my root memory now.
I've got a big problem. Earlier this afternoon I tried to unlock my screen, but the password dialog didn't appear (the background did, and I could move the pointer, but no dialog). So I restarted the computer, only for the Fedora bootup icon to get about 3/4 of the way full before the screen blanked out and I got the message "Boot has failed. Sleeping forever." I booted into the liveCD and opened the system installer to see if maybe I could just reinstall the system in place while leaving my data intact. When I got to the partitioning stage, my old partition layout was there...except one LVM volume group was totally missing. And this is the volume group that contained my / and /home, among other things. Another volume group sitting on a different RAID was still there, but ironically it was the one for short-term data.
I have three hard drives, using soft RAID and LVM. Each drive is split into 4 partitions. The first partition of each is part of a RAID-1 where /boot sits. The second of each makes up a RAID-5 on which sits my "Main" volume group for my important data (this is the one that has gone AWOL). The third of each makes up a RAID-0 on which my "Volatile" volume group sits (for /tmp, /var/tmp, and the like). The fourth is swap.
Is there any chance I can restore my volume group so my data can be recovered? I'm not sure if I've got the full layout with volume sizes written down anywhere.
i have created two physical volumes, later added volume group to it and then created logical volume and formated the logical volume n mounted it on directory now now i wanted to split the volume group but am unable to do it.If i tries it error msg displayed as existing volume group is active and i have to inactive that volume group