Software :: Moved To Mdadm Raid1 And Now Having Grub?
Nov 27, 2010
Posted this on the centos forum too, but I might get better attention here. I just moved my centos server to a mdadm raid1 array. I went partially after this guide: [URL].. What I did was to boot up a livecd and made three partitions on both of my empty disks, one for / one for swap and one for /vz (it's an openvz server). Made those partitions into seperate raid1 arrays and then rsync-ed everything from the old disk to the new partitions.
After I had moved everything I did chroot into the new raid array and edited both grub config files and fstab, according to the guide.
[Code]...
I have managed to run the system on the raid1 disks when using super grub2 disk off a cd, but it has it's own grub and can boot any distro, so I can see that the system is working fine, except for grub. I have tried installing grub both from a livecd (ubuntu 64bit) and when booted into the raid1 array, but it gives the same results as stated above.
View 2 Replies
ADVERTISEMENT
Mar 31, 2011
I've 2 servers (xen1 and xen2 - their hostnames) with perversion configuration below: Each server have 4 SATA disks, 1 Tb each.
16 Gb ddr3
debian squeeze x64 installed:
root@xen2:~# uname -a
Linux xen2 2.6.32-5-xen-amd64 #1 SMP Wed Jan 12 05:46:49 UTC 2011 x86_64 GNU/Linux
Storage configuration: Former 256 Mb + 32 Gb of 2 of 4 disks are used as raid1 devices for /boot and swap respectively. The rest of space, 970 Gb on all 4 sata disks are used as raid10. There is LVM2 installed over that raid10. Volume group is named xenlvm (that servers are expected to use as xen 4.0.1 hosts, but the story is not about Xen troubles). / , /var, /home are located on logical volumes of small size (just found out I got mixed up with lv names and partitions, but that's not the problem, I think):
[Code]...
View 3 Replies
View Related
Feb 14, 2011
I've got a couple of commercial NAS boxes and I'm wondering if they (ReadyNas duo, DLink DNS-323) or any other NAS is suitable for having their RAIDed disks moved to a software-based NAS. To be specific, I'm a big fan of the (largely) Debian-based Ubuntu. Can the aforementioned NAS drives be migrated to Ubuntu (e.g. using the mdadm Linux command)?
Secondly, is there any commercial NAS that can be migrated over? Incidentally, here is a link to somebody who succeeded in a migration:URL...My specific scenario I'd like to prepare for, is the eventual (sudden) death of one of the NAS motherboards.
View 2 Replies
View Related
Aug 11, 2010
intending to set up an all-in-one server, i threw in the ubuntu server 10.04 (amd64) cd. during the text-install, i set up the device-topology below, and it worked.
[Code]....
then i tested my raid by hot-pulling off the sda wire (ouch). worked fine, system still worked, and it also managed rebooting from the left sdb (which of course showed up being sda, lacking the first drive). now i am trying to recover this pre-crash state. adding the first disk (showing up as sdb), i can add it to md0 and let it start syncronizing for 2 hours. but... i can?t boot anymore with the recovered first disk being sda...
at first, booting got stuck in an initrd-prompt after complaining it couldn?t find my sys-logical volume. after a lot of trial and error i don?t even get any complaints, just a black screen which would let me wait for a boot for weeks... so, my system does not boot from my first disk, whether i plug in the second or not. my second disk still boots. my last attempt to get booting fine again has been: zero sda?s first and last gigabyte to kill any ids duplicate sdb?s first cylinder to sda to make it bootable reinitialize sdb?s part.table using command o in fdisk for a new disk-id recreate sda1 partition add sda1 to md0
View 1 Replies
View Related
Mar 11, 2010
I am running Kernel 2.6.18-128.el5 on a 64bit quad core machine with 8GB RAM. Using "mdadm" I setup a RAID1 array between two Western Digital 1.5TB drives. The problem is that the resync is running VERY slow. Here is a current status.
[root@royalflush shared]# cat /proc/mdstat Personalities : [raid1]
md1 : active raid1 hdc5[1] hda5[0]
1304493952 blocks [2/2] [UU]
[=>] resync = 6.2% (81592192/1304493952) finish=4280156.0min speed=4K/sec
unused devices: <none>
View 1 Replies
View Related
May 12, 2010
I'm looking to recover a RAID1 array hopefully using mdadm. Ive not really used Linux much befor but I'm keen to learn to get my data back. Basically one of the disks in my Maxtor Shared Storage II (2x500GB sata) died and I could do with either rebuilding the array or getting the data off another way.
I have a spare machine I could use for recovery process. It has a spare drive but its only 120Gig, I also have a bigger 320gig disk but thats IDE not SATA. Do I need to purchase another 500GB sata drive or can I use either of my spares? If i do need to buy a new drive could I use a 1TB or 1.5TB or will it have to be 500? Next question is what is that best version of linux to use, I have knoppix 6.2 and Ubuntu (not sure on version) already. I noticed that mdadm isn't installed by default on Ubuntu.
View 1 Replies
View Related
Jan 28, 2010
I'm trying to set up a RAID1 partition on my Ubuntu 9.10 workstation.On this dual-boot system, Ubuntu is running from a separate drive (/dev/sdc - an SSD that is quite small, which is why I need more disk space). Besides that, there are two traditional 500 GB hard drives, which have Windows 7 installed (I want to keep the Windows installation intact), and about half of the space unallocated. This space is where I want to set up a single, large RAID1 partition for Linux.
(This, to my understanding, would be software RAID, whereas the Windows partitions are on hardware RAID - I hope this isn't a problem... Edit: See Peter's comment. I guess this shouldn't be a problem since I see both drives separately on Linux.)On both disks, /dev/sda and /dev/sdb, I created, using fdisk, identical new partitions of type "Linux raid autodetect" to fill up the unallocated space.
Device Boot Start End Blocks Id System
/dev/sda1 1 10 80293+ de Dell Utility
/dev/sda2 * 11 106 768000 7 HPFS/NTFS
[code]....
But so is "Device or resource busy" when trying to create the RAID array. Quite strange.
Update: Could the device mapper have something to do with this? How do /dev/mapper and dmraid relate to all this mdadm stuff anyway? Both provide software RAID, but.. differently? Sorry for my ignorance here. Under /dev/mapper/ there are some device files that, I think, somehow match the 3 Windows RAID partitions (sd{a,b}1 through sd{a,b}3). I don't know why there are four of these arrays though.
$ ls /dev/mapper/
control isw_dgjjcdcegc_ARRAY1 isw_dgjjcdcegc_ARRAY3
isw_dgjjcdcegc_ARRAY isw_dgjjcdcegc_ARRAY2
View 2 Replies
View Related
Feb 7, 2011
I have a RAID1 array, where mdadm states that one of the disks is "removed." Naturally, I assume one of the drives has failed. The mdadm --detail command tells me that the sda drive has failed. However, further inspection from the mdadm -E /dev/sdb1 command says that sdb1 disk has been removed. I am a bit confused. Can someone clarify which drive is failed? Am I misreading the command outputs?
Code:
sudo fdisk -l
Disk /dev/sda: 500.1 GB, 500107862016 bytes
255 heads, 63 sectors/track, 60801 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
[Code]...
View 3 Replies
View Related
Nov 14, 2010
One of the hard drives in my server failed the other day, backups saved the day and downtime was only a few hours, but when setting up the new drive I went ahead and migrated to software RAID, in the hopes it may give me less downtime in the future when a drive fails. It all went rather well, but my main root partition won't finish syncing for some reason.
sda was the original drive with sda4 as /, sda1 as /boot, and sda2 as swap. sdb was the drive that failed and was replaced with the new drive. So I set up sdb with the same partitions of sda, added it to a RAID1 array, copied files from sda, and reboot to md4 as /, md1 as /boot, and md2 as swap. I added the sda partitions to the array, and the sync went off without a hitch on md1 and md2, md4 progresses well, but after a few hours /proc/mdstat just shows this:
Code:
root@d668:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md2 : active raid1 sdb2[1] sda2[0]
9767424 blocks [2/2] [UU]
[Code]...
View 3 Replies
View Related
Aug 7, 2011
I'm convinced that mdadm is going to be the death of me. I've wasted numerous hours on this so far without luck.
OpenSuse 11.4 on an old Supermicro box, creating a software RAID1 array across 2 x IDE 500GB disks. Creating /dev/md0 as a 250MB partition across /dev/sda1 and /dev/sdd1 for /boot, another 465GB partition across /dev/sda2 and /dev/sdd2 as an LVM partition to hold volumes for the various other OS filesystems. After the initial installation and configuration there were a series of mishaps with faulty IDE cables that had drives failing to show up at boot. Somehow, /dev/sdd2 got configured to array /dev/md1 as a spare drive. And nothing I've done so far gets it to show up as an active drive.
The obvious step of failing the partition, removing it, then adding (or re-adding) will bring it back as a spare. I've tried roughly a dozen different permutations of those same steps. The latest was to 'dd if=/dev/zero of=/dev/sdd2' to clear the partition. Thought this might be the trick - after the zero, mdadm -E /dev/sdd2 reported 'no superblock' and no md1 configuration.
So 'mdadm --add /dev/md1 /dev/sdd2' and it still comes back as a spare. Here is mdadm -D /dev/md1
/dev/md1:
Version : 1.0
Creation Time : Sat Jul 9 10:26:01 2011
Raid Level : raid1
Array Size : 488119160 (465.51 GiB 499.83 GB)
code....
I can't stop this array, the OS is running from there. I can't easily boot from CD to repair, all IDE ports have disks attached.
Does anyone have an incantation to promote a spare to active?
View 2 Replies
View Related
Jul 22, 2011
I have SLES10-SP3 running on an Intel SR1600URHS board with 3 hot-swap SATA disks configured using mdadm as Raid1 with hot spare. If I pull one of the active disks, all file i/o will stop for about 2.5 minutes after which it will start again and the raid array will be rebuilt using the spare disk. Is there any way I can reduce this 2.5 minutes of inactivity? I've tried setting /sys/block/sdX/device/timeout and /sys/block/sdX/device/retries to 1 for all disks, but this hasn't made any difference. The output from messages is:
12:11:56: ata2.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x2 frozen
12:11:56: ata2.00: cmd ea/00:00:00:00:00/00:00:00:00:00/a0 tag 0 cdb 0x1e data 0
12:11:56: res 40/00:03:00:00:20/00:00:00:00:00/b0 Emask 0x4 (timeout)
[code]....
View 1 Replies
View Related
Feb 27, 2011
I've faced the problem with server freeze on heavy write.
System
CentOS 5.5 x64_86 with latest updates and kernel (2.6.18-194.32.1). Also tried 2.6.18-194.26.1 and 2.6.37-2 from ELRepo with the same results.
CPU: Intel(R) Core(TM) i7 CPU 920 @ 2.67GHz
Memory: 3 x 2Gb DDR3.
HDDs: 2 x Western Digital WDC WD1002FBYS-02A6B0
[Code].....
View 19 Replies
View Related
Jun 27, 2009
I have a used but good harddrive which I'd like to use as a replacement for a removed harddrive in existing raid1 array. mdadm --detail /dev/md00 0 0 -1 removed1 8 17 1 active sync /dev/sdb1I thought I needed to mark the removed drive as failed but I cannot get mdadm set it to "failed". I issue mdadm --manage /dev/md0 --fail /dev/sda1But mdadm response is:mdadm: hot remove failed for /dev/sda1: no such device or addressI thought I must mark the failed drive as "failed" to prevent raid1 from trying to mirror in wrong direction when I install my used-but-good disk. I want to reformat the good used drive first right? I believe I must prevent raid array from automatically try to mirror in the wrong direction.
View 7 Replies
View Related
Dec 7, 2010
Just finished installing F14 with RAID1 setup for 2 hdd (SATA''s). Entire drives are mirrored, including SWAP. As I had done in the past, I was planning on installing grub on MBR of 2nd hdd. In prep for this I did the following to locate the grub setup files:
grub> find /grub/stage1
find /grub/stage1
(hd0,1)
[code]....
I was surprised, expected to get (hd0,0) & (hd1,0), not (hd0,1) & (hd1,1)
running "fdisk -l" I get:
Disk /dev/sda: 120.0 GB, 120034123776 bytes
255 heads, 63 sectors/track, 14593 cylinders, total 234441648 sectors
Units = sectors of 1 * 512 = 512 bytes
[code]....
MBR is the first 512 bytes of the drive. Each partition has a boot sector. In my case grub stage 1 is on 2nd partition of sda & 2nd partition of sdb. What i dont understand is how grub stage 1 can be on sda2 & sdb2, since I am assuming that sda1 & sdb1 would be the first partitions of the drives & therefore contain the MBR's. Maybe this might be because sda1 & sdb1 are SWAP partitions?
View 5 Replies
View Related
May 5, 2010
I have an Ubuntu Server on (8.10) running under Citrix XenServer (though that shouldn't make a difference).
I installed on a single disk:
xvda1 - 200 MB - /boot
xvda2 - 9.8 GB - LVM (ubuntu-base)
The LVM is:
swap - 1.0 GB
root - 8.8 GB - /
I have successfully gotten this converted to RAID1 by adding a new drive (xvdb) and following the Debian howtoforge article [URL]
What I have not been able to do, is get grub working properly.
If I fail xvdb and reboot the system, everything comes up and I can reboot and run.
If I fail xvda and reboot the system, XenServer gives me a bootloader error. i.e.: no grub
If someone has done this, can they tell me what grub commands to run to get a successful boot of the primary disk fails?
View 1 Replies
View Related
Jul 26, 2014
I have a VIA Epia M 5000 system with 2 western digital 1TB NAS SATA drives connected through SATA<->IDE adapter. Everything installs and writes as expected except... grub. It never boots, after a message 'Grub loading' I always get 'error: no such disk'. I've tried numerous times and has been attemping to fix the issue for the past 2 days.
/dev/sda
0.999TB 0xfd linux raid autodetect partition
1GB 0xfd linux raid autodetect, logical partition
/dev/sdb exactly the same
0.999TB 0xfd linux raid autodetect partition
1GB 0xfd linux raid autodetect
/dev/md0
RAID1 of /dev/sda1 and /dev/sdb1
marked as ext4, boot point /
/dev/md1
RAID1 of /dev/sda5 and /dev/sdb5
marked as swap
update-grub2 in rescue mode generates grub.cfg with SET root=/mduuid/UUID_OF_SDA1
then after that there's search --no-floppy etc --set root=/mduuid/UUID_OF_MD0
I'm writing this from memory but simply the two uuids are different. Is this correct? I get those UUIDs to compare from blkid. All partitions are marked as bootable. grub-install /dev/sda and grub-install /dev/sdb produces no errors. grub-install /dev/md0 does not work, complains about superblocks or something similar.
Grub.cfg file contains insmod raid mdraid1x and similar lines, so that should be ok. Grub drops to rescue mode with message error: no such disk. Not device, but disk. Google finds many results for 'no such device' error, but I am not getting that error. 'ls' produces
(hd0) (hd0,msdos1) (hd1) (hd1,msdos1) ls ANY_VALID_PATH produces empty newline being printed, nothing more.
setting prefixes manually does not work, with error message 'error: file not found'. ls (hd0,msdos1)/boot/grub also produces empty line being printed. Rescue CD and auto-assemble of md0 and md1 works, the files are there, everything okay, except grub.
View 3 Replies
View Related
May 3, 2011
Followed steps 2-5 and purged/reinstalled grub now it boots as it should, NO idea where it was messed up.[URL].. I had 9.10 running in raid1 and upgraded my hardware (cpu, mb, memory etc) and wanted to do a fresh install of 10.04 to get updated. After following the various guides online such as [URL]...It begins to load grub and drops to a "grub> " shell. Which I have to do the following to get it to boot.
Code:
set root=(md1)
linux /vmlinuz root=/dev/md1 ro
initrd /initrd.img
boot
Then it boots up normally and I can use it like any other desktop. I've been over my grub.cfg and /etc/defaut/grub files and cannot find the issue. At this point I'm wondering if the fact it's a raid1 setup is keeping grub from finding it's files.
[Code]...
View 9 Replies
View Related
Sep 19, 2014
I am running a 14 disk RAID 6 on mdadm behind 2 LSI SAS2008's in JBOD mode (no HW raid) on Debian 7 in BIOS legacy mode.
Grub2 is dropping to a rescue shell complaining that "no such device" exists for "mduuid/b1c40379914e5d18dddb893b4dc5a28f".
Output from mdadm:
Code: Select all # mdadm -D /dev/md0
/dev/md0:
Version : 1.2
Creation Time : Wed Nov 7 17:06:02 2012
Raid Level : raid6
Array Size : 35160446976 (33531.62 GiB 36004.30 GB)
Used Dev Size : 2930037248 (2794.30 GiB 3000.36 GB)
Raid Devices : 14
[Code] ....
Output from blkid:
Code: Select all # blkid
/dev/md0: UUID="2c61b08d-cb1f-4c2c-8ce0-eaea15af32fb" TYPE="xfs"
/dev/md/0: UUID="2c61b08d-cb1f-4c2c-8ce0-eaea15af32fb" TYPE="xfs"
/dev/sdd2: UUID="b1c40379-914e-5d18-dddb-893b4dc5a28f" UUID_SUB="09a00673-c9c1-dc15-b792-f0226016a8a6" LABEL="media:0" TYPE="linux_raid_member"
[Code] ....
The UUID for md0 is `2c61b08d-cb1f-4c2c-8ce0-eaea15af32fb` so I do not understand why grub insists on looking for `b1c40379914e5d18dddb893b4dc5a28f`.
**Here is the output from `bootinfoscript` 0.61. This contains alot of detailed information, and I couldn't find anything wrong with any of it: [URL] .....
During the grub rescue an `ls` shows the member disks and also shows `(md/0)` but if I try an `ls (md/0)` I get an unknown disk error. Trying an `ls` on any member device results in unknown filesystem. The filesystem on the md0 is XFS, and I assume the unknown filesystem is normal if its trying to read an individual disk instead of md0.
I have come close to losing my mind over this, I've tried uninstalling and reinstalling grub numerous times, `update-initramfs -u -k all` numerous times, `update-grub` numerous times, `grub-install` numerous times to all member disks without error, etc.
I even tried manually editing `grub.cfg` to replace all instances of `mduuid/b1c40379914e5d18dddb893b4dc5a28f` with `(md/0)` and then re-install grub, but the exact same error of no such device mduuid/b1c40379914e5d18dddb893b4dc5a28f still happened.
[URL] ....
One thing I noticed is it is only showing half the disks. I am not sure if this matters or is important or not, but one theory would be because there are two LSI cards physically in the machine.
This last screenshot was shown after I specifically altered grub.cfg to replace all instances of `mduuid/b1c40379914e5d18dddb893b4dc5a28f` with `mduuid/2c61b08d-cb1f-4c2c-8ce0-eaea15af32fb` and then re-ran grub-install on all member drives. Where it is getting this old b1c* address I have no clue.
I even tried installing a SATA drive on /dev/sda, outside of the array, and installing grub on it and booting from it. Still, same identical error.
View 14 Replies
View Related
Nov 22, 2009
Here's a brief description of my system:
120GB Sata HDD - Primary OS drive
3 x 1.0TB Sata HDD - Raid 5 array
This is on a C2D MSI P35 Platinum board. Anyway, did a fresh install of F12 on the 120GB, which I had problems with - Anaconda refused to see the drive. Fedora Live could see it fine, and it was listed as an 'nvidia_raid_member' - no idea why, but I completely erased the disc under the Live CD and proceeded to install F12.
Once F12 was installed, I loaded up mdadm to re-activate my Raid 5 array, using 'sudo mdadm --assemble --uuidthe uuid) - and it started with only 2 of the 3 drives. My /dev/sdb drive did not activate into the array, due to what mdadm said was a mismatched UUID. Ok, so I erased /dev/sdb, intending to rebuild the array. Erased /dev/sdb, and then attempted 'sudo mdadm --add /dev/md0 /dev/sdb' and I get this error: "mdadm: Cannot add disks to a 'member' array, perform this operation on the parent container" - I can find NO information on this error message.
[Code].....
I don't believe the hard drives are connected in the exact same order they were in before - I disconnected everything in the system and blew it out (it was pretty dusty)
View 1 Replies
View Related
Jul 18, 2011
I have a raid5 on 10 disk, 750gb and it have worked fine with grub for a long time with ubuntu 10.04 lts. A couple of days ago I added a disk to the raid, growd it and then resized it.. BUT, I started the resize-process on a terminal on another computer, and after some time my girlfriend powered down that computer!
So the resize process cancelled in the middle and i couldn't acess any of the HDDs so I rebooted the server.
Now the problem, the system is not booting up, simple black with a blinking line. Used a rescue CD to boot it up, finised the resize-process and the raid seems to be working fine so I tried to boot normal again. Same problem. Rescue cd, updated grub, got several errors: error: unsupported RAID version: 0.91. I have tried to purge grub, grub-pc, grub commmon, removed /boot/grub and installed grub again. Same problem.
I have tried to erased mbr (# dd if=/dev/null of=/dev/sdX bs=446 count=1) on sda (ide disk, system), sdb (sata, new raid disk). Same problem. Removed and reinstalled ubuntu 11.04 and is now getting error: no such device: (hdd id). Again tried to reinstall grub on both sda and sdb, no luck. update-grub is still generating error about raid id 0.91 and is back on a blinking line on normal boot. When you'r resizeing a raid MDADM changed the ID from 0.90 to 0.91 to prevent something that happend happened. But since I have completed the resize-process MDADM have indeed changed the ID back to 0.90 on all disks.
I have also tried to follow a howto on a similar problem with a patch on [URL] But I cant compile, various error about dpkg. So my problem is, I cant get grub to work. It just gives me a blinking line and unsupported RAID version: 0.91.
View 2 Replies
View Related
Jan 11, 2010
I am planning on setting up a 4x1TB RAID5 with mdadm under Ubuntu 9.10. I tried installing mdadm using "sudo apt-get install mdadm", all worked fine except for the following error: Code: Generating array device nodes... /var/lib/dpkg/info/mdadm.postinst: 170: /dev/MAKEDEV: not found failed. The end result is the /dev/md0 device has not been created, as can be seen here:
Code: windsok@beer:~$ mdadm --detail /dev/md0 mdadm: cannot open /dev/md0: No such file or directory After googling, I found the following bug which describes the issue: [URL] However it was reported way back in April 2009, and it does not look like it will be fixed any time soon, so I was wondering if anyone knows a workaround for this bug, to get me up and running?
View 4 Replies
View Related
May 9, 2011
In ubuntu Natty with the notification pop ups like for when the song changes in Banshee, they are in my top right corner. Is there anyway to move them?
View 3 Replies
View Related
Jul 20, 2011
I had a Thinkpad T61 running Debian 6 that died. I got a T400 and had hoped to save my debian 6 installation by putting the original T61 hard drive into the T400. When I do the machine gets to the Grub menu, then it starts booting and the text starts moving up the screen. Then it goes to a small blinking cursor at the top left corner and stays there.The T61 had an Nvidia card, the T400 has the switchable Intel/ATI. I went in the BIOS and told it to boot using the Intel only but no go. I'm thinking that Grub needs configging to see my new hard drive but I don't know how to do that.
View 14 Replies
View Related
Mar 23, 2010
I moved an LVM to a new computer, but everytime I reboot I have to do vgchange -ay and then mount it. I tried putting it in fstab, but it seems it doesn't work because the volume group is disabled when booted
View 1 Replies
View Related
Jan 14, 2010
I hooked up my computer to a TV via VGA and after I was done my icons moved around on the taskbar. Usually the 'kevin' was on the far right. Is there a way to change it back?
View 1 Replies
View Related
Feb 21, 2010
why a file can't be moved to trash while it can be deleted forever!! does any of you have a clue?
View 1 Replies
View Related
May 10, 2010
I have a script I only want to run when a window is being dragged around. Is there an environmental variable or something that will tell me whether one is moving and which one it is? I'm using Ubuntu Lucid if it matters.
View 9 Replies
View Related
Apr 7, 2011
I have recently moved. I was on a comcast Internet system and had the Server working there. I have moved to where I am on a different server. I have been issued a static IP address. I have configured my Linksys server to reserve an address for my Ubuntu Web Server. When I try to get to my web site the connection times out not getting there. I remembered I had to forward some services, ports over to the reserved IP address for the server. I can't for the life of me remember the ports I forwarded. how to get this forwarding done. Or what else might be keeping me from seeing my webpate - [URL]?
View 3 Replies
View Related
May 16, 2011
ubuntu 10.04 Easy Peasy eeepc 900I have messed up my home directory and can not get out of the messI wanted to move the /home to be on the 16gb partition instead of the default which was on the 4gb with the installation. I copied/saved /home to a temporary directory and then moved the contents of /home to the 16gb disk and mounted it on /home At this stage all seemed to be working correctly and applications worked as before. However when I rebooted it can not find my home directory, complains a lot, and if you bypass the messages ends up with a minimal GUI screen which does not respond to anything I tried. It seems that the 16gb disk is not being mounted.
View 4 Replies
View Related
Jun 18, 2010
I'm running Ubuntu 9.10 AMD64 with KVM and several VMs running the same OS. Everything was running fine and the VMs were using the host's network via bridging. The decision was made to move the server to completely different network. I shut it down, the system was moved and hooked up, and I got it up on the new network pretty easily--just modified /etc/network/interfaces and /etc/resolv.conf. No networking problems at all.
Then I fired up virt-manager, powered up one of the VMs, and made the same kinds of changes. It's not talking beyond the VM server. I can ping the host, but can't ping the gateway. "tcpdump -i eth0" shows plenty of traffic.
View 4 Replies
View Related