Fedora :: Keep Getting "A Hard Disk Is Failing" Warnings From The Disk Utility
Jun 8, 2011
I am on F15 32-bit with GNOME 3. I keep getting "A Hard Disk Is Failing" warnings from the Disk Utility, very frequently. Is this a serious issue? Because I knew this to be a bug in Palimpsest DU back in F13/14. Also how can I disable any notifications from this application?
Last week, i updated, my fedora. After that, during every boot up, I am getting a warning message like " Your hard disk may failing". It indicates that it is due to bad sectors. But I don't think so. There was a bug reported for a similar problem in fedora 11. I think it is not fixed yet. Hard disk is not having any problem during data access. Other OS including windows are not giving any warning message.
How to diagnose a possibly failing harddisk. This is the system disk of a Windows XP machine that will not boot. It gets to the "progress bar" that runs across the screen a few times but then it always stops in the same spot. If I boot from a live Ubuntu CD what tools should I use to find out if the disk itself is failing? The disk is under warranty but if I take it to the shop and the problem turns out to be software errors then I have to pay so I want to be sure what is going on before going there..
I have a 2 TB disk in an external SATA dock, formatted with a single ext3 (Linux) partition, which doesn't show up in the Windows 7 Computer Management->Disk Management utility, even as a raw/blank disk. I've verified that there's nothing wrong with the disk by connecting it to my Linux machine and mounting it, and I've verified that the dock is functioning properly by connecting a different FAT32-formatted disk, which mounts flawlessly as expected.I realize that I can't actually read the ext3 partition without additional software (e.g., Ext3IFS), but why doesn't the disk show up at all? Is there some sort of stupid anti-Linux filter built in? Is there any way to force Windows to recognize the disk, so that I can at the very least use direct block access with it?
Background: I want to clone an identical 2 TB disk onto this one. Due to my hardware layout, it's much easier to have the source disk attached to one machine and the destination disk connected to another, and do the clone over the network (the network is not a bottleneck with switched gigabit ethernet), than it is to hook them both up to one machine.(1) I did this once before when both machines were running Linux, but I've since upgraded the destination machine and decided to switch back to Windows for regular desktop use. I've got Cygwin installed, and have verified that the same basic method (dd + nc) will work, but I can't do anything if Windows doesn't even consider the destination disk to exist.I only have one eSATA port on each machine. Opening them up just to do this clone is a rather large annoyance. Also, since this is my backup disk, I'd like to eventually automate the cloning from the active disk to another one that I regularly swap with a third disk that I store off-site.
I recently tried Fedora on my laptop (previously Debian; I was bored one day) and gnome-disk-utility (palimpsest) warned me that my hard drive had numerous bad sectors. I re-installed Debian to find that this software was installed before so why had it not warned me?
When I load the disk utility, it says SMART is not available. I've got smartmontools installed, I can run a self-test with smartctl but I don't think this shows bad sectors. I've tried starting smartd on startup but the disk utility never changes from "SMART is not available". It is possible for it to work with this hardware as it works in Fedora on this laptop; any ideas?
I don't understand disk sizes in Linux. I have a 500GB drive. It's ext4. I have run "tune2fs -m 0" on it to reserve the amount of space reserved for root to 0.
I'm using Ubuntu 10.04 that comes with a Disk Utility. When I run "System->Administration->Disk Utility (palimpsest)" the disk shows up as 500GB (see picture). But when I run df -h it shows up as 459GB. So, I don't understand the discrepancy.
When I run df I get the following:
Question: Why is Disk Utility showing me something different than "df"?
I had this corrupted external hdd and so I formatted the main partition on it on windows but messed up in the formatting and ended up having to format the entire thing. I got some weird message about it not being initialized (no not mounted) so I was in compmgmt.msc in windows and right clicked it in device manager and it asked for master boot or GUID I selected the latter and formatted. Worked fine and all for a bit but now it doesn't show up as a drive. I noticed when using compmgmt.msc it showed up that it had installed driver software and was being recognized but in the partition editing area there was nothing on this drive, reinstalling driver software doesn't seem to help. Also GParted wont load up when I have it plugged in and Disk Utility doesn't show it. I am requesting help to fix this problem within Ubuntu 10.10 somehow so I can use it properly.
after installing Ubuntu on one WD 500 GB hard disk and after making mistake and pasting wrong code into Terminal:my OTHER WD 500 GB hard disk that was also in the system (I guess it was "hd1") - died.The problem must be, I guess, I typed wrong code: "hd1,1" instead of "hd0,0".)500 GB (NTFS) of data was on that other (non-Ubuntu) hard disk, and now I can not access it anymore. While booting, system gives "Hard Disk Error" warning and stops.One again: I installed Ubuntu od one hard disk and at the end of instalation I pasted wrong code for GRUB, giving address of another hard disk. Now that other hard disk has error and will not work
I have a sata 320 gb with mandriva linux 2009.1 on it.And it is what curently atached to my cpu. It is shown as 'sda' in the partition table.I also have another 40gb hard disk with windows xp installed on it.It is shown as 'hda' in the partition table . Now what i want to do is attach this 40gb hard disk to my pc and configure grub on my 320gb hard disk('sda') so as to boot windows xp(which is residing on the second hard disk,'hda')Can anyone tell me if what im doing is feasible or not? If it is feasible,can anyone suggest me how to get it working. I know i just need to add 2-3 lines to my grub.conf, but dont know what exactly i need to write.
I had a dual boot (windows 7 + debian), both of them installed in my internal hard disk, with the GRUB in it. I have recently installed a second linux distro (mint), but I put it in an external hard disk. Now the GRUB allows me to boot any of the three operating systems, but I need the external disk to do it. It seems that after the mint installation the GRUB is now working from the external disk (if the external disk is not connected, the machine does not boot.) �Is there a way to change the location of the GRUB, to the internal hard disk of my laptop?
I was using Terminal and browsing a directory in my home folder. My "home" directory is located on "/dev/sdb1". When in Terminal I typed "ls" in one of my directories and the output was garbage. The output didn't show the files in the directory. I think it said something like, "input/output error". Unfortunately, I didn't write the exact error down. Instead I rebooted.The hard disk with the problem is:
Code: $ sudo hdparm -I /dev/sdb [sudo] password for brian:
I created a thread about a problem a I had with my hard disk clicking whilst idle little while ago and I may now have stumbled upon a possible solution. The strange thing with the problem is that Ubuntu/Kubuntu didn't cause this problem but Opensuse 11.2 does.
I installed Fedora 13 to have a glimpse of what all the fuss was about and noticed that I had the same problem (hard disk clicking whilst idle ~ every 20 secs or so). Now there's a wiki on this subject and a few bug reports: [url]
Some ATA harddrives perform very frequent head unloads under Linux significantly shortening their lifespans. Root cause
The inactivity timer for head unload is configured too aggressively either via ATA APM (Advanced Power Management) feature or other non-standard means. Such aggressive settings are very fragile to changes in IO pattern and under Linux many such drives unload their heads only to re-load them shortly. Note that this relentless unloading/reloading cycle can also be triggered under Windows by installing programs which can alter the IO pattern (e.g. certain vaccine programs which runs in background).
Now two of the listed models with this problem are basically identical to my model (Dell Inspiron 1520) and basically share the same hardware: Dell Vostro 1500 and XPS 1520.
The workaround listed is to:
set APM to 254
Furthermore, there is a script: Storage-Fixup which can also be downloaded from opensuse software search. Indeed there is a report of this for a Vostro 1500: Gmane Loom
The report suggests looking at: Disk Power Management - openSUSE which lists a method to create a configuration file to management disk power management:
My question is whether I could download the storage-fixup rpm [url] has a description of it and it can be found: Software.openSUSE.org) and install it to (hopefully) solve the issue or should I follow the method given in: Disk Power Management - openSUSE
ran out of space in my /home dir. Have a second hard drive to install and would like to designate it as additional space for /home. I do not want to mount it as a dir inside my home I would like it to simply work as though my /home simply has more space available to it.
I'm trying to resize a partition on an IDE hard disk to use the entire disk but can not get more than a 309GB partition. I can get 295, 300, 301, 302GB, etc... fine but start getting problems with anything over 309GB. I get the following error with 310GB or more:
error: block relocator should have relocated 533 Warning: You should reinstall your boot loader before rebooting. Read section 4 of the Parted User documentation for more information. I am using Slackware 12.1, GNU parted 1.8.8, ext2 filesystem.
I'm working on setting up a new NAS. I installed Karmic desktop on a 160 GB HD using the default settings.
Now I've added three 1TB drives and want to make them a RAID-5 array with LVM on that, and 1 ext4 partition. I want to use LVM so I can add drives and expand the array later.
So far I've been using Disk Utility (Palimpsest Disk Utility) and it's been great! A wonderful addition to Karmic! I got the RAID-5 array setup with no problems using disk utility. So now I have a 2000 GB raid-5 array setup in Disk Utility and I need to get LVM setup.
Problem is: I don't see any sign of LVM in Disk Utility. I've been googling all night and I can't find any documentation for setting up LVM in Disk Utility, just people saying that it's supported.
I tried installing the lvm2 package, rebooting, and then looking around again. No luck.
So, what am I missing? Should there be LVM options in Disk Utility? Where is it? Is there a better/easier way to configure lvm?
point me in the direction to get a step by step guide to setting up a Raid 5 using the Disk Utility and 3 spare drives? I have the main OS files on a 80gig drive and I would like to mount the 3 drives as Raid 5.Just shooting in the dark now.. Screen shot is attached. [URL]...
Has anybody ever used Disk Utility to set up software RAID? Here I am running terminal commands (I'm a terminal junkie) and I just happen to stumble across instructions that indicate "Or you can just set it up through Disk Utility."
Sure enough in disk utility, it looks like all of the configurable options are there. It makes me wonder, though... is this kind of GUI functionality something that isn't really solid? Or does it operate predictably and effectively?
I have a 160GB harddrive which I installed a F12, would like to upgrade to a bigger drive, but I hate to have to re-install everything.
Recommend a good disk copy utility? The utility should be able to not only copy files, but boot sector and everything. So I just need to make a copy, change my BIOS to boot from the new drive and run everything as before.
When I use disk utility to expand my RAID array it creates a partition on my 1.5TB drive which it would like to add to the RAID 5.
However, none of the drive existing on the RAID are partitioned so what I think has happened is the partition itself has created a difference of about 2 million bytes smaller than the others and thus unable to add the component.
How can I specify the exact bytes for my hard drive partition so that I can add this to the array?