General :: Extend The Lvm And The Filesystem To The Second Disk?
Dec 18, 2010
I've added a second drive to a system and I need to extend the lvm and the filesystem to the second disk. Is there a way to do this online with centos 5.5? I specifically need extending the actual ext3 filesystem which seems to be the trick part.
i'm using RHEL 5.0 with ext2 or ext3 filesystem, now i wish to use ext4 without loose this Opreating System.... or without using another distro of linux...
I wonder if this is possible to extend or regrow the Linux hard disk partition from 8 GB to 20 GB without losing the existing data on the partition ?at the moment this Ubuntu Linux is deployed on top of VMware and I've just regrow the hard drive from 8 GB into 20 GB but can't see the effect immediately.can anyone suggest how to do this without losing the data ?
I've 80G hard-disk with dual boot ( XP with 5.5G and Ubuntu 10.10 with 4.5G ). After recent updates there is only 1G of space left on Ubuntu.I've 3 more drives with around 25G left and want to extend Ubuntu by another 10G.How can I do that?
I wanted to extend my raid array with one disk, but I made a major error. I forgot partition the new disk to utilize the full 640GB. I used the following commands to extend the array:
I'd like to make the ls -laR /media/myfs on Linux as fast as possible. I'll have 1 million files on the filesystem, 2TB of total file size, and some directories containing as much as 10000 files. Which filesystem should I use and how should I configure it?As far as I understand, the reason why ls -laR is slow because it has to stat(2) each inode (i.e. 1 million stat(2)s), and since inodes are distributed randomly on the disk, each stat(2) needs one disk seek.Here are some solutions I had in mind, none of which I am satisfied with:Create the filesystem on an SSD, because the seek operations on SSDs are fast. This wouldn't work, because a 2TB SSD doesn't exist, or it's prohibitively expensive.
Create a filesystem which spans on two block devices: an SSD and a disk; the disk contains file data, and the SSD contains all the metadata (including directory entries, inodes and POSIX extended attributes). Is there a filesystem which supports this? Would it survive a system crash (power outage)?Use find /media/myfs on ext2, ext3 or ext4, instead of ls -laR /media/myfs, because the former can the advantage of the d_type field (see in the getdents(2) man page), so it doesn't have to stat. Unfortunately, this doesn't meet my requirements, because I need all file sizes as well, which find /media/myfs doesn't print.Use a filesystem, such as VFAT, which stores inodes in the directory entries. I'd love this one, but VFAT is not reliable and flexible enough for me, and I don't know of any other filesystem which does that. Do you? Of course, storing inodes in the directory entries wouldn't work for files with a link count more than 1, but that's not a problem since I have only a few dozen such files in my use case.
Adjust some settings in /proc or sysctl so that inodes are locked to system memory forever. This would not speed up the first ls -laR /media/myfs, but it would make all subsequent invocations amazingly fast. How can I do this? I don't like this idea, because it doesn't speed up the first invocation, which currently takes 30 minutes. Also I'd like to lock the POSIX extended attributes in memory as well. What do I have to do for that?Use a filesystem which has an online defragmentation tool, which can be instructed to relocate inodes to the the beginning of the block device. Once the relocation is done, I can run dd if=/dev/sdb of=/dev/null bs=1M count=256 to get the beginning of the block device fetched to the kernel in-memory cache without seeking, and then the stat(2) operations would be fast, because they read from the cache. Is there a way to lock those inodes and/or blocks into memory once they have been read? Which filesystem has such a defragmentation tool?
I've just bought a new SSD hard drive:Kingston SSDNow V-Series SNV125-S2/128GB 2.5'' 128GB SATA/300The question is which filesystem whould you recommand and why?BTRFS vs NILFS2 or EXT4?If you choose ext4 would you enable jurnalling?I'm very close to choose Btrfs.Any experience with running any of these on your SSD?
Currently we use an iSCSI SAN as storage for several VMware ESXi servers. I am investigating the use of an NFS target on a Linux server for additional virtual machines. I am also open to the idea of using an alternative operating system (like OpenSolaris) if it will provide significant advantages.What Linux-based filesystem favours very large contiguous files (like VMware's disk images)? Alternatively, how have people found ZFS on OpenSolaris for this kind of workload?
Background: My mother's HP laptop had Ubuntu and Vista on it, Ubuntu my brother's doing. He decided he wanted to take off Ubuntu yesterday (he had forgotten the password), and deleted the partition that it was contained within. The computer now boots to this error.
Inventory: We no longer have the install disk for Windows Vista, he cannot tell me what version he used of Ubuntu, what partition it was on, any of the specs for the machine, or generally any information about the system. All I am aware of is that error on the boot-up screen. I have nothing else to work with.
I would like to remove Grub, and Ubuntu, and leave Windows intact (the request of the owner of the computer), but I have no idea what commands I could use to get rid of either when I can't access Windows, or how to properly remove them if I did access Windows.
I've had a look at some similar threads but as I'm very new to linux they're already a bit technical for me. Sorry, this calls for someone with patience. I gather from other threads that disconnecting an external drive without unmounting is a no-no, and this seems to be the likely cause. Now the disk is read only and I'm unable to change any settings through the usual control panel on ubuntu. I'm just not familiar with the terminal instructions. I tried to cut and past a few command lines from other threads but I got some warnings that proceding could damage data. Like this one: WARNING! Running e2fsck on a mounted filesystem may cause SEVERE filesystem damage.
why my df -h command is telling me I'm missing 18G?
*** Filesystem Size Used Avail Use% Mounted on /dev/mapper/vg_aikifedora-lv_home 357G 272M 339G 1% /home ***
This was 'df -h /home' straight after a fresh install. It says it's 357G in size, but even though 272M is in use, it's telling me I have only 339G left available. Where did that extra 18G go?
I've setup a filesystem on a RAID 0+1 and am looking at moving root filesystem from a single disk to the new one. I could not install CentOS on mirrored filesystem as the RAID card did not have a pre-built driver for CentOS 5.3, so I had to compile the driver after installing the system.What I'm going to do now is:
1. Mount the new mirrored filesystem under /root1 2. use find | cpio to copy everything from the existing / to /root1 3. use grub to create a boot record on /root1 4. edit /root1/etc/fstab to point / to the new disk 5. reboot the system and keep my fingers crossed
Ive setup a filesystem on a RAID 0+1 and am looking at moving root filesystem from a single disk to the new one. I could not install CentOS on mirrored filesystem as the RAID card did not have a pre-built driver for CentOS 5.3, so I had to compile the driver after installing the system.
What Im going to do now is:
1.Mount the new mirrored filesystem under /root1 2.use find | cpio to copy everything from the existing / to /root1 3.use grub to create a boot record on /root1 4.edit /root1/etc/fstab to point / to the new disk 5. reboot the system and keep my fingers crossed
I know I have done this command on out Red Hat systems. But for some reason in CentOS its giving me problems. I am trying to add more space to / as its at 100% full. Not sure what filled up the 4G is assigned to it but I did load Virtualbox and a widowsxp system.
Here is what i have:
df Filesystem 1K-blocks Used Available Use% Mounted on /dev/hdc3 4061572 4050416 0 100% / /dev/mapper/rootvg-rvg_datalv 253934980 313572 240514208 1% /data
I have just formatted an external USB disk with a JFS filesystem. The partition shows up in 'Computer', and it mounts, but if I try and copy and files onto it, it will not do it. Clearly, Nautilus is mounting it read-only. How do I get this to behave like my USB Flash drive, where I plug it in, and its automatically mounted read/write?
I'm trying to use xrandr to extend the desktop on my laptop to my lcd monitor. I have tried a few different ways but no matter what I do the desktop will end up on my LCD and the laptop is the extended desktop. Anyone know where I am going wrong? Here is the command I've been using:
What I need is a program that will extend a X-windows display over a network, similar to nvidia's TwinView (which I am already using on the main machine for its 2 current monitors).However, I cannot use Xdmx (as suggested by all the threads I found), as I would like it not to require me to manually start the x server and viewer, and have all the machines must be running a *nix OS, which won't work for me, as the secondary machine is running windows.I would also like it to hopefully use VNC to share the desktop, as I would be connecting to it from a windows machine over ethernet.By "extend" I mean one that would, say, for example, add another X display with a given resolution and position and serve that over VNC. I have yet to find any programs that will do this.
I am trying to access a Magneto Optical disk written under OS/2 hpfs filesystem. I cannot locate a dang thing about support of CentOS for OS/2. I found a module for Linux OS/2 support that dated back to 1993, have no idea how to compile it, used alien to create an RPM package, said compiled succesfully, still can't mount the drive (/dev/sdb) with filesystem type -t hpfs. Can anyone out there help me, or does CentOS just not do this?
I recently installed Bio-Linux 5.0 as a dual boot system with XP for some bioinformatics applications, but Im having some problems with the amount of disk space which can be allocated specifically for the Ubuntu install.
Ive been using blastclust to analyse some very large data sets, which keeps on crashing due to filesystem running out of disk space.
When I installed Bio-Linux 5.0 from the live cd, the maximum size I could allocate to the install was 30 GiB, and I havent been able to find a way to change this.
Ive tried using System->Administration->Partition Editor using the live cd, and can view / delete the partitions, but I cant find a way to specifically alter the disk space allocation for Ubuntu.
How do I increase the filesystem size to larger than the current 30 GiB?
I have been having problems with filesystem corruption on my eeepc 1000H for a long time now. I have tried using different filesystems, kernels and distributions (arch, slackware) to no effect. I am starting to grow suspicious that this problem lies somewhere else, as I haven't seen anyone else having similar problems in such a variety of scenarios.
I have tried testing my ram using memtest86+, didn't come up with anything after a full run through. I also have tried using e2fsck -c to check for bad blocks, it finds none. I had a go at using smartctl but wasn't really sure what I was doing. I did a long test and it came up with nothing anyway.
This problem is in addition to the problems I've been having with my intel graphics chip and KMS. A lot of the time there are lockups when booting into X, which can only be gotten out of by a hard reset. This is sometimes what causes the original filesystem errors. I've stopped messing around with KMS for now to eliminate this but my current system in unbootable. I'm guessing my disk is wrecked but have as yet seen no definitive proof. Can anyone recommend anything that I should do?
I am currently on ext4 with a custom kernel 2.6.33-rc6 (the stock kernel shipping with slackware does not have the elantech extension for psmouse included). When I was using arch, I was just using the stock kernels.
I am very new to linux, and I have a question regarding the filesystem check (fsck). The power recently went out and when I tried to restart linux the following error appears:
*/dev/sda1 contains file system w/errors, check forced it then goes on to say..
*An error occured during the file system check. Dropping you to a shell; the system will reboot when you leave the shell. Give root password for maintenance (or type Control-D to continue) I wasn't sure what to do, but checked some other online forums and they suggested running fsck manually - so I typed in the root password - and used the command, "fsck -A -V ; echo == $? ==" it then gave the following message
*WARNING!!! Running e2fsck on a mounted filesystem may cause SEVERE filesystem damage *Would you like to continue (y/n)
Again, I wasn't sure what to do so i just checked no. I then manually turned off the computer and was prompted at the beginning to press Alt-3. I was brought to another screen and it informed me one of the drives was degraded and suggested rebuilding the array. I tried doing this, but it still brings me back to the original error of, "/dev/sda1 contains file system w/errors, check forced," and the process continues.
Also, when I tried to rebuild the array, I didn't backup any of the data on our home directory before doing this (which was probably a big mistake). After being prompted to type the root password, I was able to give the ls command and look at all the directories...the home directory where our data was stored was empty and I am afraid I may have lost some information. Is there a possibility that data was lost when I was trying to rebuild using the old drives?