Fedora :: Managed To Reduce The Logical Volume To Free Some Space But Cant Seem To Reduce The Physical Volume
Jan 1, 2010
so i have f12 installed on my hd with lvm using the whole extent of the HD , i want to reduce it so i can dual boot it with a windows system, i managed to reduce the logical volume to free some space, but i cant seem to reduce the physical volume, is this possible and how ?
I have one HD and a VG spanning it's entirety. I resized a LV and freed up 10GB within the VG, but I want the 10GB outside the boundary of the VG so I can install another OS for testing purposes. For some reason I'm not able to do this. I don't know if I understand LVMs correctly. Maybe there can be only one VG on a HD?
My laptop has 2Gb of ram, 4Gb of swap, 40Gb hdd and an intel pentium 1.40GHz cpu, even when compiling stuff or maxing out everything the 4Gb of swap is never touched, with such a small hdd I'd like to reduce the swap to about 2Gb (just in case) to free up some space, does anyone know what commands/tools are available to accomplish this?
My computer: (Lenovo T61 Thinkpad, running fc11 for about 2 and half months). Apparently I when I made my partitions I didn't leave quite enough room in my root directory, because I just completely ran out. Here is how my hard drive is partitioned:
The root had about 15 gigs on it, which just filled up. When I restarted to see if that would help, when it rebooted it went fine up to the log-in screen. Instead of the usual fedora blue background, it was black except for the log-in window, which looked very low-res. A little pop-up kept coming up saying the GNOME power configuration settings failed to load or something. When I logged in, the whole screen was black except for the mouse, and I could get no response. I have plenty of space left in home, so I rebooted to rescue mode using the first fedora installation disk, and tried the following command:
lvreduce -L90G /dev/mapper/DRIVE
which only returned:
lvreduce: relocation error: lvreduce: symbol dm_tree_node_size_changed, version Base not defined in file libdevmapper.so.1.02 So I couldn't reduce the size of home, and thus couldn't increase the size of root.
a) the lack of memory in root the probable cause for my computer not working
b) there a good way to reduce home and increase root while running this live disk
Note: When I am looking at it now in the logical volume manager, it says that on the whole physical volume there is only 400MB free. However, when I last looked (about 30 mins before I started having problems) it said there were about 100 Gb free.
Edit: Nevermind. I did some more research and it turned out to be more of a gnome power manager thing rather than a memory space thing, although I'm certainly going to increase my root memory now.
I have a 7.9 TB logical volume I've created from 8 1 TB RAID 0 devices. The volume is formatted with XFS so I can resize when ready. However, I think I want to do something that is not possible. I have 2.5 TB free on my logical volume. I'd like to shrink the volume down to be 6 TB by getting rid of 2 of the 1 TB devices in the physical volume. However pvmove seems to require free extents in order to work. Do I need to add 6 TB of storage, pvmove everything onto it, and then decommission the original 8 1 TB physical devices from the volume?
How to create multiple Logical Groups out of a single Physical Volume? Here is the Physical Volume I have created:
Code: # pvdisplay --- Physical volume --- PV Name /dev/sda9 VG Name myVG1 PV Size 54.88 MB / not usable 2.88 MB Allocatable yes PE Size (KByte) 4096 Total PE 13 Free PE 11 Allocated PE 2 PV UUID bon4Ao-vmgC-aP1h-EC9X-w3tN-YXNu-0N2dAw
This is how I am creating a Logical Group out of the above Physical Volume:
Code: # vgcreate myVG1 -s 4m /dev/sda9 Display:
Code: # vgdisplay --- Volume group --- VG Name myVG1 System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 5 VG Access read/write VG Status resizable MAX LV 0 Cur LV 2 Open LV 1 Max PV 0 Cur PV 1 Act PV 1 VG Size 52.00 MB PE Size 4.00 MB Total PE 13 Alloc PE / Size 2 / 8.00 MB Free PE / Size 11 / 44.00 MB VG UUID O6ljYC-bflz-EUTd-nf34-8gYe-Fh39-Bh3cOg
But I am unable to create one more Logical Group out of this Physical Volume. Can we accomplish it? Or do we always extend our current Logical Group to utilize the available space of a Physical Volume?
how easy it would be to read the contents of a physical disk that was part of a larger logical volume. The disk contains a "Linux LVM" partition that spans its entire size. My problem is that one of my disks died, and I have to send it back for a warranty replacement. However, the disk is dead, and I can't zero it out. I'm just trying to assess how difficult it would be (or at least how likely it would be) for a tech that's checking out the disk to get at the data.
I've got a question on free disk space. I'm currently running CentOS 5.5 on in Xenserver virtual environment. We've had an issue with disk space. My question is as follows: - from a ssh connection i run df -h this gives the value of 90% used leaving me with 9GB. If I use system monitor via a VNC connection the free disk space value is 20GB free on the same volume. Which one is correct? I do use SNMP to monitor the same volume and should alert me when < 10% is free I know this works as I set the alert threshold to < 90% I get an alert.
I just read and learned about logical volume management today. I have a server running RHEL5.4, LVM2. I have 1 physical volume, with one volume group, and 3 logical volumes. I have no free extents, nor do I have any in my volume group (not sure if it's possible to have free in one and not the other anyway), and I am running out of space on one of my logical volumes. Doing a df -h shows 96% of 9.7GB used on /dev/mapper/MainVG-root, mounted at /. So here's the stupid question: how can I find out what directories/files are taking up what space within this logical volume? As I said I have 3 all together, and the other 2 are mapped to /var and a /var pgsql sub-directory. I figured I could get the sizes of the other directories under / and drill down accordingly, but I seem to be missing some basic rule because the commands I am using and the values I am getting don't add up.
For example, it seemed logical to me to do an ls -lsh on / to try and identify the largest directories. Each directory is listed as being ~4-8K in size. That doesn't make sense to me. So I decided to do a du -sh on each directory. Having done this on all of the / sub-directories and added up those values, there is not enough reported usage here to equal 8.9GB of used space (as df -h / reports).how they would find out how the 9.7GB here is being allocated? Preferably without scripts as I am not ready to add a layer of complexity to this yet without understanding some fundamentals.
I have a rack of four 1TB drives all partitioned identically with three primary partitions. On each drive
- the first partition is only 64MB; - the second is a large 900GB partition and - the last holds all the remaining space
mdadm has been used to set up /dev/md0 - RAID1, comprised of /dev/sda1 and /dev/sdb1 /dev/md1 - RAID5, comprised of /dev/sda2, /dev/sdb2, /dev/sdc2, /dev/sdd2 /dev/md2 - RAID5, comprised of /dev/sda3, /dev/sdb3, /dev/sdc3, /dev/sdd3
OK, so it was a silly mistake to make - but I am now need to increase the size of /dev/md0. My thinking is to reduce the size of md1 so that I can grow md0.
On md1 I have two logical volumes. I've successfully reduced the size of the volume so that I can reduce the size of md1. Now I'm at the nervous stage; I can find little written on the topic of shrinking RAID5 arrays - and even if I do this I'm unsure if I can move partitions around to regain the space I so desire.
HPDL385 with dual raid controllers (8 disks each). During the install of the ISO, it sees the raid controllers individually. I tried "One Generic Drive" but it still only partitions one of the raid controllers.Is it possible after the OS is installed to configure the space as one logical volume?
I'm rearranging a bunch of disks on my server at home and I find myself in the position of wanting to move a bunch of LVM logical volumes to another volume group. Is there a simple way to do this? I saw mention of a cplv command but this seems to be either old or not something that was ever available for linux.
I've added a new LUN to my Centos 5.2 server using Powerpath and have added it to an ext3 logical volume. I extended the logical volume using lvextend and the new space shows up correctly in lvdisplay. What I'm having problems with is getting Centos to see the new disk space (df -h shows 500GB, not 600GB as expected). I've tried running a resize2fs on the new volume but it tells me that "the filesystem is already n blocks long. Nothing to do". Does any one know where I'm going wrong? If possible I'd like to sort this without a reboot.
I have a system with a 2TB RAID level 1 installed (2 x 2TB drives, configured as RAID1 through the BIOS). I installed Centos 5.5 and it runs fine. I now added another 2x2TB drives and configured them as RAID1 through the BIOS.
How do I add this new RAID volume to the existing logical volume?
I have a 500 GB hard drive connected via a SATA/USB adapter. After connecting it, I did a basic format of the drive with no problem. I then attempted the "Wipe Free Space" option under the File menu in BleachBit, selecting the external drive, and it began the process. It gives me a time estimate of about 300 minutes, but, instead, after a few minutes it says "Done" and gives me the message "[Errno 27] File too large. Disk space recovered: 0 Files recovered: 0 Errors: 1"
What happened? What do I need to do to get this volume wiped?
I have a Fedora 8 system that uses LVM on one of it's drives (/dev/sdb2). One of the logical volumes is getting full (LogVol02). There is an unused, unmounted logical volume (LogVol03) available. I can see two possible options.
1) Mount the unused logical volume (LogVol03) on a new mount point (/home2) and create more space there 2) Delete the LogVol03 logical volume and extend the nearly full volume (LogVol02) into the now available space.
Option 2 seems like the better approach, since it will seem seamless to the system users. I'm looking for suggestions on how I should go about doing this and what I need to look out for. Is it better to use the command line tools (lvm ...) or the GUI (system-config-lvm) to do this?
i have a 700GB ext4 partition for storage purposes. By default it has 5% of the space (35GB!) reserved for root, which does not make sense for this partition. how can i reduce this percentage? there is already a lot of data on the partition and i'm afraid that mke2fs would erase all the data. is there a way to change the percentage without touching the data?
So I got a bad physical volume inisde my logical volume. I want to do this safely rather than tinkering around, how can I get the bad physical disk out and look at the data on the other 2 drives to see if I can save anything? Its just the standard fedora setup where it combines all the disks, nothing fancy.
I have the volume group activated as a partial, and now I just want to see the data on the other sections how could I mount that?
I've got a big problem. Earlier this afternoon I tried to unlock my screen, but the password dialog didn't appear (the background did, and I could move the pointer, but no dialog). So I restarted the computer, only for the Fedora bootup icon to get about 3/4 of the way full before the screen blanked out and I got the message "Boot has failed. Sleeping forever." I booted into the liveCD and opened the system installer to see if maybe I could just reinstall the system in place while leaving my data intact. When I got to the partitioning stage, my old partition layout was there...except one LVM volume group was totally missing. And this is the volume group that contained my / and /home, among other things. Another volume group sitting on a different RAID was still there, but ironically it was the one for short-term data.
I have three hard drives, using soft RAID and LVM. Each drive is split into 4 partitions. The first partition of each is part of a RAID-1 where /boot sits. The second of each makes up a RAID-5 on which sits my "Main" volume group for my important data (this is the one that has gone AWOL). The third of each makes up a RAID-0 on which my "Volatile" volume group sits (for /tmp, /var/tmp, and the like). The fourth is swap.
Is there any chance I can restore my volume group so my data can be recovered? I'm not sure if I've got the full layout with volume sizes written down anywhere.
There are 2 volumes on single group. The boot partion is a physical volume and the system is a logical volume. The disk has more room up to 40GB. How can I extend the logical volume. Tried system-config-lvm, but it does not gives the option.
I (again) ran into trouble since two months i was trying out fedora lovelock "nightly build" in a virtualbox my host is Ubuntu 11.04 (previously 10.10 i upgraded yesterday) (had worked fine .. just was slow because i alloted only 512 MB ram) ... now i uninstalled virtualbox and deleted .virtualbox in home folder, but i couldn't recover my 8 GB space when i installed i made virtual disk fixed storage (my mistake i guess!! ) ) i don't know much about it either .. just wanted to try out fedora 15 so experimented it using virtuallbox ...
i'm running out of disk space (don't have bucks to purchase external hard disk .. otherwise i wouldn't have cared .. also don't have time to do a reinstallation of ubuntu/some other os (may be fedora) on my laptop ...
i just got 24 GB left .. i intend to store a few more movies and songs .. i would run out of disk space fairly quickly ... 8GB would be buffer if i get it back ...
I'd like to clone a partition, and then restore it to a logical volume. I have all three operating systems at my disposal (Mac, Windows, Linux Live CD) What is the best way to achieve this. The partition I am trying to resize is only 200MB, so I can store it on usb if need be.
Ok so I have one drive. /boot /lv_root and /lv_swap
At the end of the drive I have 32 gigs of free space still contained in the logical volume group. I want to remove it from the LVG but this is on one device. Supposedly there is a way to do this, pvresize and fdisk.
Originally Posted by source
#I've tried to shrink the PV with pvresize which didn't throw errors -
#but fdisk still shows me the same LVM partition size as before.
That's normal. pvresize "just" updates the PV header and VG metadata.
#So I guess the partition table has to be modified somehow?
Yes. That was mentioned in my reply: "Then shrink the partition in the partition table."
You can use fdisk or any other partition table editor for this. Some don't support resizing a partition. In that case, you can delete and create a smaller one. If doing the delete/create dance, you *must* create the new partition on the same cylinder boundary as the current one to preserve the current data.
Ive read from every source on LVM its not possible to do this. Why on earth would any Linux developer put LVM on a single drive system by default? Were they even paying attention? I dont mean to go off on a rant but if there are multiple drives LVM makes sense. However if you only have one large drive LVM holds your system hostage and you have to crawl thru the pit of hell to get it back.
I understand you have a choice in the matter when you install Fedora but its really the worst possible choice for default. Many newcomers to Linux run into this problem with LVM. If you cannot resize LVG's the software should have never been put into a Linux distro in the first place.
I have a Fedora 15/Ubuntu 11.04 dual boot. In Fedora, I just created a new logical volume. How do I ensure that anything I put in it - all files and folders - will be accessible in Ubuntu? II also want to be able to add files and folders in Ubuntu and then read and write them in Fedora...
Here's the output from "lvdisplay" for the logical volume in question: --- Logical volume --- LV Name /dev/fedora/Everything VG Name fedora
/home/gwyn/Everything is the mount point for the logical volume. I got full access over all files in the LV.
But when I shutdown Fedora and start up Ubuntu, I have to run the same chmod command again. How can I make read/write permissions endure between both distros?
I just installed linux fedora 14 on my hp probook 4320s with installation CD with this name: Fedora-14-x86_64-Live-Desktop. Then I installed it on the hard disk. During installation I chose to encrypt hard disk. When I try to access my hard disk it says "unable to mount 250GB LVM2 Physical Volume, Not a mountable file system". What can I do to get access to my hard disk?
I don't know much about lvm and I've managed to screw up a drive. I had a 500GB drive with FC14 on it and I wanted to copy over a load of data to my new 1TB that was replacing it. I set up my new install the same way as the old...including the same volume names (error number 1 I think) I successfully mounted the old/500GB drive (using vgscan and vgchange -a y etc.) using a laptop (running FC13) and an external hdd cradle. I could access the files I wanted but this wasn't the machine I wanted to copy them to (I was doing this while waiting for the install to finish on the new drive).
When I tried the same process on the new install I found that having two lvm with the same name meant I couldn't mount the external one. So I opened the disk utility (palimsest) and was going to change the name of the old volume group but it wouldn't let me do that. I then thought maybe I could get away with just changing the name of the partition where the files were and maybe I could add it to the mounted group or something so I changed it to lv_home2. This changed the name of my new/1TB lv_home to lv_home2 as well. So thinking that wasn't the answer I just changed the name of the new lv_home2 back to lv_home.
From that point on I haven't been able to see the old drives partitions (the new volume group still works so far). I has a physical volume but the volume group and volume names are gone from view. When I try to vgscan on my main computer or the laptop I had it working on earlier I get: