I have a 2 TB disk in an external SATA dock, formatted with a single ext3 (Linux) partition, which doesn't show up in the Windows 7 Computer Management->Disk Management utility, even as a raw/blank disk. I've verified that there's nothing wrong with the disk by connecting it to my Linux machine and mounting it, and I've verified that the dock is functioning properly by connecting a different FAT32-formatted disk, which mounts flawlessly as expected.I realize that I can't actually read the ext3 partition without additional software (e.g., Ext3IFS), but why doesn't the disk show up at all? Is there some sort of stupid anti-Linux filter built in? Is there any way to force Windows to recognize the disk, so that I can at the very least use direct block access with it?
Background: I want to clone an identical 2 TB disk onto this one. Due to my hardware layout, it's much easier to have the source disk attached to one machine and the destination disk connected to another, and do the clone over the network (the network is not a bottleneck with switched gigabit ethernet), than it is to hook them both up to one machine.(1) I did this once before when both machines were running Linux, but I've since upgraded the destination machine and decided to switch back to Windows for regular desktop use. I've got Cygwin installed, and have verified that the same basic method (dd + nc) will work, but I can't do anything if Windows doesn't even consider the destination disk to exist.I only have one eSATA port on each machine. Opening them up just to do this clone is a rather large annoyance. Also, since this is my backup disk, I'd like to eventually automate the cloning from the active disk to another one that I regularly swap with a third disk that I store off-site.
OS: Debian unstable 32bit, kernel 2.6.32-2, grub 1.98 from late january 2010 (only have working net-access from work now, so I am grabbing information from memory). EXT3 and EXT4 support is compiled into the kernel along with chipset/scsi/sata support (not as modules), and I have tested to boot ext3 with it before proceeding. Prereq: my old disk started to have too much S.M.A.R.T errors, so I bought another one, put in a USB cabinet, added swap and ext4 partition/filesystem to it, and copied over all data from the old system to the new that was mounted at /dest using the command "find ./ -xdev -print0 | cpio -paV0 /dest". Swiched disks, so I now have the ext4 disk sitting at /dev/sda (partitions: sda1 => ext4, sda2 => swap), and booted into rescue-mode from cdrom, using /dev/sda1 as root with a shell on. After doing this, I performed the following commands:
mount --bind /dev /dest/dev chroot /dest
modified the /etc/default/grub to instruct the kernel to boot using ext4, ran grub-install --recheck /dev/sda ran update-grub to modify /boot/grub/grub.cfg (which looks as it should) After doing this, grub finds my partition and mounts it. It however stalls with the message: "warning: unable to open an initial console" and does nothing after this point. I have no ramdisk, but my old kernel booted fine from ext3 (and still does if I copy it to a ext3 partition), and since the ext4 support is compiled into the kernel - should I really need a ramdisk?
I don't understand disk sizes in Linux. I have a 500GB drive. It's ext4. I have run "tune2fs -m 0" on it to reserve the amount of space reserved for root to 0.
I'm using Ubuntu 10.04 that comes with a Disk Utility. When I run "System->Administration->Disk Utility (palimpsest)" the disk shows up as 500GB (see picture). But when I run df -h it shows up as 459GB. So, I don't understand the discrepancy.
When I run df I get the following:
Question: Why is Disk Utility showing me something different than "df"?
I have windows and linux distros on my /dev/sda. I tried to install fedora 13, but after reboot I cannot boot up to any of installed systems. I'm getting: Non-System disk or disk errorreplace and striky any key when readyWhen I use ubuntu 10, boot from first hard disk, I'm able to get GNU GRUB version 1.98...nd boot up to any system. but without ubuntu in a cd-rom I'm getting this error.I tried: grub-install /dev/sdX in one of my installed linux distros, but without any success.
S.M.A.R.T.tells me my primary fedora drive has lots of bad blocks,so I ordered a used identical drive from ebay.I want to do the very basic process of making a full backup of my existing drive - so as to replace it with a more or less identical replacement. Logically identical - not necewsarily sector to sector identical.im not literate in unix and its children - and seem to find a plethora of descriptinos of how to do what I want to do. which is
1. format the drive - what you may call low level format - ie op sys independent.
2. create a partition system compatible with my existing disk.
3. copy everything from my existing disk to the new disk.
4. put existing disk on the shelf as a backup and start to use new disk.
I had a dual boot (windows 7 + debian), both of them installed in my internal hard disk, with the GRUB in it. I have recently installed a second linux distro (mint), but I put it in an external hard disk. Now the GRUB allows me to boot any of the three operating systems, but I need the external disk to do it. It seems that after the mint installation the GRUB is now working from the external disk (if the external disk is not connected, the machine does not boot.) �Is there a way to change the location of the GRUB, to the internal hard disk of my laptop?
My Ubuntu system is occasionally becoming very sluggish. I'm running many things simultaneously and it's very difficult to tell which program is the culprit.
I suspect that the sluggishness is due to disk activity since the CPU usage is consistently under 50% on each of the 4 cores of the CPU, and over 30% of the 6GB of RAM are free.
Is there a tool that can show me in real time the number of disk IO operations per second and the amount of data read/written per second? Can all this info be broken down and displayed per process?
Some thing is using up a huge amount of my disk space about 10G and I can not determine what it is. When I look at my disk usage in system monitor it say I have used about 25G and when I scan the directory in disk usage analyzer the entire file system used is 15G.
This is the third 9.10 install to do this on two different laptops, so wondering what's up...
In both cases, the goal was to leave a large chunk of unpartitioned disk after the Ubuntu partitions, for a second OS install or a filesystem Ubuntu cannot create like NTFS.
When I install with manual partitions, the system can't boot and asks for me to insert a system disk and press any key. When I reinstall telling Ubuntu to "use the entire disk" it then works.
First laptop, first try:
Remainder of the 500GB disk is free space.
Fails to boot, "insert system disk".
First laptop, second try without the /boot partition:
Remainder of the 500GB disk is free space.
Fails to boot, "insert system disk".
"use entire disk" works perfectly.
Second laptop, first try:
Same thing, non-system disk or disk error, insert system disk.
Second try "use entire disk" is currently in progress but I expect the same to happen.
I have an SiI hardware SATA RAID card, with two 500GB disks in mirrored RAID configuration. When I first plugged them in and set it up, things seemed to work ok, but on boot the raid controller told me that the RAID needed rebuilding, and it would happen automatically after POST. So I didn't worry about it, and the drive mounted fine, and it's been that way for years. I just went in and manually on-line rebuilt the RAID in the controller's BIOS, and now when I boot into Ubuntu, both disks show up in fdisk, but neither show up in /dev/disk/by-uuid. Am I missing something?
I downloaded the latest version of wubi and when I click to run i get the error "pyrun.exe - No Disk. There is no disk in the drive. insert a disk into drive DeviceHarddisk2DR2".
ran out of space in my /home dir. Have a second hard drive to install and would like to designate it as additional space for /home. I do not want to mount it as a dir inside my home I would like it to simply work as though my /home simply has more space available to it.
WinXp sp3 is on disk sdb, then installed Ubuntu 10.04 on sda, can go into diff OS without any problem. I am going to move sda to another machine, when I unplug sda, WinXp can't start to boot on sdb. How to fix it?below is my case output$ sudo fdisk -l
Disk /dev/sda: 160.0 GB ... Device Boot Start End Blocks Id System
i have amd athlon x2 3gb RAM ddr2 two hdd one 80gb other 1tb in this computer i got only ubuntu all disk in EXT4.i want to copy 10gb many file of 700mb or 1gb ( many linux ISO )the transfer is 1mb/s this is SOOO SLOW.i did a touch /forcefsck just in case but nothing happen it is still slow!!what might be the problem ??
I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.
I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).
These are the commands I used:
Quote:
p -a -d -R -v -t /media/raid_array /b* cp -a -d -R -v -t /media/raid_array /d* cp -a -d -R -v -t /media/raid_array /e* cp -a -d -R -v -t /media/raid_array /h*
[Code]....
I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...
So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.
It's still trying to use /dev/disk/by-uuid/412d... to boot.
What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.
I have a netbook I'm not using and which I transformed into a server with Apache, Tomcat6, Netatalk, Webmin, BIND9 and Tor.
Problem is, the disks never stop spinning because all of the programs write a few kb at least every few seconds to disk, even when nobody is connected to it.
My question is: Is there a way to have the computer boot from disk like normal (maybe even a squashfs), keep ALL CHANGES to ram and then save to disk when either the ram is full (unlikely because the server is rebooted every few days) or at shutdown?
I thought about a mixture of ramfs and unionfs but I'm not good enough yet...
my Fedora 11 system is not starting anylonger. It stops with the message:
Code:
VFS: Can't find ext4 filesystem on dev dm-0
The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).
I received the following error when I got home from work today. If this was a windows environment, my first inclination would be to boot off my dvd and then run a chkdsk on the drive to flag any bad sectors that might exist. But there's a complication for me.
Code: Select allThis message was generated by the smartd daemon running on: host name: LinuxDesktop DNS domain: [Empty]
The following warning/error was logged by the smartd daemon: Device: /dev/sdc [SAT], 1 Currently unreadable (pending) sectors Device info: WDC WD5000AAKS-65V0A0, S/N:WD-WCAWF2422464, WWN:5-0014ee-157c5db9a, FW:05.01D05, 500 GB For details see host's SYSLOG.
You can also use the smartctl utility for further investigation.The original message about this issue was sent at Sun Feb 14 13:43:17 2016 MST.Another message will be sent in 24 hours if the problem persists.
From gnome-disks Code: Select allDisk is OK, 418 bad sectors (28° C / 82° F)
I did a bit of reading and it seems that most people suggest using badblocks to first get a list of badblocks from the drive and save it to a file. Then use e2fsck to then mark the blocks listed in the badblocks file as bad on the hard drive. My problem here is that this drive is part of a RAID5 array that hosts my OS. I wanted to confirm if this was still the correct process.I boot to my Live Debian disk, stop the raid array if it's active. Then run badblocks + e2fsck commands on the drive in question and then reboot.
Is it possible to install GRUB in the MBR of the only bootable disk in the system, but load configuration and images from another disk?Basically I want to install GRUB on /dev/sda, but menu and images will be under /dev/sdb2.Note: /dev/sdb is not bootable.
There is a disk 500 gb, it is broken on /boot and on /root and on /dev/sda1 and /dev/sda2. Whether prompt it is possible to redistribute a disk without loss of data namely it is necessary to make/boot and two equivalent on disk volume.
I created a thread about a problem a I had with my hard disk clicking whilst idle little while ago and I may now have stumbled upon a possible solution. The strange thing with the problem is that Ubuntu/Kubuntu didn't cause this problem but Opensuse 11.2 does.
I installed Fedora 13 to have a glimpse of what all the fuss was about and noticed that I had the same problem (hard disk clicking whilst idle ~ every 20 secs or so). Now there's a wiki on this subject and a few bug reports: [url]
Problem Description
Some ATA harddrives perform very frequent head unloads under Linux significantly shortening their lifespans. Root cause
The inactivity timer for head unload is configured too aggressively either via ATA APM (Advanced Power Management) feature or other non-standard means. Such aggressive settings are very fragile to changes in IO pattern and under Linux many such drives unload their heads only to re-load them shortly. Note that this relentless unloading/reloading cycle can also be triggered under Windows by installing programs which can alter the IO pattern (e.g. certain vaccine programs which runs in background).
Now two of the listed models with this problem are basically identical to my model (Dell Inspiron 1520) and basically share the same hardware: Dell Vostro 1500 and XPS 1520.
The workaround listed is to:
set APM to 254
Furthermore, there is a script: Storage-Fixup which can also be downloaded from opensuse software search. Indeed there is a report of this for a Vostro 1500: Gmane Loom
The report suggests looking at: Disk Power Management - openSUSE which lists a method to create a configuration file to management disk power management:
My question is whether I could download the storage-fixup rpm [url] has a description of it and it can be found: Software.openSUSE.org) and install it to (hopefully) solve the issue or should I follow the method given in: Disk Power Management - openSUSE
I installed Debian on my PC with a Acer Stock motherboard (xc600) with amd64 and after the installation finished it told me to remove my installation media and reboot. After reboot I was returned this message ' ERROR: No boot disk has been detected or the disk has failed.'. I have verified with gparted using mint live OS that I have Debian installed on my system.
I got believes that this may have be caused by a broken grub or I need to configure something I don't know how in BIOS.
I will update the topic later..
My installation media was a USB 2.0 flashdrive with a Debian 8.2 Jessie Installer and 9 different Linux distros. I have installed Debian multiple times before on my laptop and never had this problem so I know how to go through the installation process and set the partitions.