Red Hat :: Increasing Disk Device To / Root Disk Enterprise Server 4?
Mar 25, 2010
I am having problem with the root disk / fill up. So now that i have the disk information provided from SAN team Id 0:3, how would i do to add this device to / to increase space on Redhat Enterprise server 4?
I'm running out of space in wubi. Online wubi help didn't help much since they suggest creating extra virtual disk space(similar to having a diffrent partition i guess) .None of them speak about increasing the size of /root disk space(or root.disk). I store all files in space shared with windows or external disk and use ubuntu only to install and use softwares and browsing. So how do increase the available space for installing more softwares?
I have installed ubuntu 10.10 inside windows (windows 7) from the ubuntu home edition CD.I have allocated a disk space of 12GB. How could I increase the disk space to 20 GB without reinstalling ubuntu?
I am running OpenSUSE 11.4, and have 2 partition in it, one is / (about 10GB), another one is /home (about 50GB). I usually put into sleep when I'm away from my computer. It had been few days I never shut down my computer, and today I got a warning message mention that my disk space (/home partition) is full. I check my disk space in Dolphin's properties menu for the /home directory, found out that it only used up 10GB disk space. I did a check on the "My Computer" on the desktop, the status is showing full usage (100%) in red color. I did df -h command, the partition for the /home is showing 100% used as well. I don't really know what is going on, and then I restart my PC. It back to normal after I come back to my Linux, which is 10GB disk space used. I don't know whether this is a bug in OpenSUSE or not.
I recently installed Bio-Linux 5.0 as a dual boot system with XP for some bioinformatics applications, but Im having some problems with the amount of disk space which can be allocated specifically for the Ubuntu install.
Ive been using blastclust to analyse some very large data sets, which keeps on crashing due to filesystem running out of disk space.
When I installed Bio-Linux 5.0 from the live cd, the maximum size I could allocate to the install was 30 GiB, and I havent been able to find a way to change this.
Ive tried using System->Administration->Partition Editor using the live cd, and can view / delete the partitions, but I cant find a way to specifically alter the disk space allocation for Ubuntu.
How do I increase the filesystem size to larger than the current 30 GiB?
There isn't any separate file system for /home and we have only one (/) root file system for everything else on the system. Is there any way that we can still implement quotas for users through their home directories was mounted on (/) root file system. Do we need to have a separate file system (/home) compulsory for implementing disk quotas?
I have some servers running rhel5,they are in mirror disk 2 disk there is no any hardware RAID only software RAID and they in remote place.
Assume RHEL5 hda(Disk1) & hdb(Disk2) currently in mirror, i have to install some patches in it. Before that i have to break the mirror and boot from hdb if it is successful. then i have to apply the patches to hda.
After that i have to boot from hda and check whether the latest patches updated and working fine. finally i have put them in mirror. the same scenario for solaris.
Is there any C function that will translate UUIDs into device names? I have a little graphical mount tool that can read user-mountable device names from /etc/fstab and lets you cycle through the list and mount or unmount them. But it doesn't work with UUIDs, which are preferred these days. Is there any way around this?
I have a 4 * 1.5TB RAID5 disk array (software linux RAID, formatted with jfs) on my Fedora 12 system and want to expand it by adding another 1.5TB disk. I have added a drive to the system and conducted a simple performance check on it to make sure it was functioning properly:
Code:
# dd if=/tmp/bigfile.dat of=/dev/sdg1 5478774+1 records in 5478774+1 records out 2805132609 bytes (2.8 GB) copied, 168.77 s, 16.6 MB/s
But 16.6 MB/s is lousy. I ran an iostat -dmx 2 on this drive at the time of this lousy performance, and typical output was:
(note that sda and sdb are a linux raid mirror set for the / filesystem that holds /tmp). I formatted the new drive (/dev/sdg1) with jfs and mounted it under /mnt2:
Code:
# jfs_mkfs /dev/sdg1 jfs_mkfs version 1.1.13, 17-Jul-2008 Warning! All data on device /dev/sdg1 will be lost! Continue? (Y/N) y
[code].....
I would like to know what the performance will be like before I add the disk to the array because I don't want to wait for the whole array to be rebuilt before finding out my array is performing badly. The array is used as part of a mythtv system and has up to 6 simultaneous recordings running on it, so it needs to perform well.
I have a Corsair R60 ssd disk which is a disk with both sata and usb connectors. But the usb thing seems to be a bit non-standard, or maybe its just my fedora linux.When I insert the disk using a usb cabel to a running Fedora 14 linux system, a device called /dev/sg3 is added but that is all. No new /dev/sd* device is created so I can't mount the disk.
So the disk is there. (The last entry) but my linux will for some reason not see it as a usb hard disk. When I insert other usb disks they work fine. It is only this specific disk which causes problems. I have tried on 3 different computers with the same result.
A hint to the problem may be that if I add the disk to a windows system(With usb) the disk is called "A fixed disk" and not a portable disk as expected. The disk works fine with linux If i connect it with the sata cabel, but I would really like to have it working with usb too. (To mount it on computers without sata).
Added:I did try to mount /dev/sg3 but mount say that its not a block device. (File say Its a character special device).
Added output from dmesg:
[ 97.454073] usb 7-1: USB disconnect, address 2 [ 105.913055] hub 2-0:1.0: unable to enumerate USB device on port 3 [ 107.048054] usb 2-3: new high speed USB device using ehci_hcd and address 5 [ 107.162900] usb 2-3: New USB device found, idVendor=1b1c, idProduct=1ab8
[code]....
I found an other guy with exactly the same problem [URL] so I think its beginning to look like a bug in the drives firmware or in the linux kernel.
Final update:Corsair have said that the disk design is broken and there does not seem to be any way to make it work.
2 days ago i have tried to install fedora 11 ...but after i did achieved installation successfully ...and rebooted system ,the device hasn't booted from hard disk ,alternatively it has booted from the network (Broadcom)- the third booting choice in bios - . i have tried three different DVDs ....but no thing changes ...!
Sometimes when I restart, my NTFS drives will mount in incorrect directories. It seems to only happen when I plug in USB devices such as flash drives, and keep them in when I boot. I have the fstab file configured correctly, but it still resorts to some odd default mounting points. Edit suposedly the device name changes whenever I boot with a flash drive plugged in. Is there anyway to mount a disk to a dir without pointing to its changing device name.
After 2-3 partition an extended partition automatically created in which I am not able to create specified capacity i.e., say I want 150g of /photos partition, the /videos partition is automatically reduced and a free space at the end appears. Some free space is always there which i am not able to understand. Nevertheless i clicked to create, but I get an error viz. 'device not created'.
Running RHEL 5 in a VMWare image, needed to add a second virtual disk, which worked fine - added it in VMWare, used fdisk and mkfs to format and mounted it as /SI50 so I could then access it. Did all this as root of course. Now I have created a new user and when I log in not only do I not see the new drive, but if I try and mount the drive I get an error "only root can do that". I absolutely have to see this drive as this new user so mounting it only for root to see is no good at all. I'm sure I'm fundamentally doing something wrong here, because obviously the way drives appear in Linux is completely different to windows.
With the continued growth of SATA devices, I have really been having problems with my boot loader. In the old days, it was pretty easy to make sure that the IDE devices stayed put. However, now everything is an sd device. Specifying a root directory of /dev/sda1 in lilo is no longer a sure thing. If you have a USB device plugged in or an ESATA device, your sd devices are no longer guaranteed to stay put.In reading through the lilo documentation, it appears that you can specify root devices by UUID or even by LABEL.
Some thing is using up a huge amount of my disk space about 10G and I can not determine what it is. When I look at my disk usage in system monitor it say I have used about 25G and when I scan the directory in disk usage analyzer the entire file system used is 15G.
my Fedora 11 system is not starting anylonger. It stops with the message:
Code:
VFS: Can't find ext4 filesystem on dev dm-0
The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).
I've mounted Ubuntu 10.04 ISO using Daemon Tools and Virtual CloneDrive, but I end with thisI've disabled my DVD drive in device manager, but I still get it no matter how many times I try.
I am developing for a Linux based device for which the HOT PLUG option is deactivated. As part of optimizing the code, we also don't want to create device files for unused devices. We understand that both USB attached and fixec SCSI hard disks would create device files like /dev/sda,/dev/sda1 /dev/sdb, /dev/sdb1 etc. Is this understanding correct?
In the case of USB attached SCSI devices, would driver create this device file entry? How is it created? Can somebody please tell me how it is being created automatically. In case I attach a fixed SCSI hard disk before boot up(and create device file /dev/sda1), would USB SCSI device driver create device files starting from /dev/sdb, automatically.
For backup purposes, I built a RAID6 array on USB. When during the night the backup filesystem is being mounted, every so many days, a random disk fails:
Code: sd 9:0:0:0: [sdb] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE sd 9:0:0:0: [sdb] Sense Key : Illegal Request [current] sd 9:0:0:0: [sdb] Add. Sense: Invalid field in cdb[code].....
In other words, my RAID set fails every so many mounts, but I know how to fix it.What I want to find out is:
- Why is a random drive failed every now and then?
- How can I prevent that drive from being failed?
BTW: I have a second RAID set that has been functioning for years without error, so the setup I use must be correct. The only difference between both RAID arrays is the different vendor of the disks.
I have installed vmware in windows server 2003 enterprise edition. When I installed I changed the default installation folder from c: drive to i: drive, as I do not have space in c: drive. After I captured red hat linux cd in vmware and reached partitioning section and continued I got a message like "an error has occured no valid devices were find on which to create new file systems, please check the hardware for the cause of the problem". The machine is IBM server....
I received the following error when I got home from work today. If this was a windows environment, my first inclination would be to boot off my dvd and then run a chkdsk on the drive to flag any bad sectors that might exist. But there's a complication for me.
Code: Select allThis message was generated by the smartd daemon running on: host name: LinuxDesktop DNS domain: [Empty]
The following warning/error was logged by the smartd daemon: Device: /dev/sdc [SAT], 1 Currently unreadable (pending) sectors Device info: WDC WD5000AAKS-65V0A0, S/N:WD-WCAWF2422464, WWN:5-0014ee-157c5db9a, FW:05.01D05, 500 GB For details see host's SYSLOG.
You can also use the smartctl utility for further investigation.The original message about this issue was sent at Sun Feb 14 13:43:17 2016 MST.Another message will be sent in 24 hours if the problem persists.
From gnome-disks Code: Select allDisk is OK, 418 bad sectors (28° C / 82° F)
I did a bit of reading and it seems that most people suggest using badblocks to first get a list of badblocks from the drive and save it to a file. Then use e2fsck to then mark the blocks listed in the badblocks file as bad on the hard drive. My problem here is that this drive is part of a RAID5 array that hosts my OS. I wanted to confirm if this was still the correct process.I boot to my Live Debian disk, stop the raid array if it's active. Then run badblocks + e2fsck commands on the drive in question and then reboot.