Today I made a samba-share out of /Video/Rorschach to easily put files in there from my windows7-machine (the plan is to steam from my CentOS-server to my HTPC which hasn't arrived yet).I started to put movies in there. It went just fine for a while but then I got this message:
How is that even possible when df -h looks like this?:
[root@Rorschach Rorschach]# df -h
Filesystem Size Used Avail Use% Mounted on
446G 70G 354G 17% /
I recently followed this guide to create a RAID1 [URL]... First I partitioned the disks with fdisk. I made the RAID array with
mdadm --create /dev/md0 --level=1 --raid-devices=2 /dev/sdb1 /dev/sdc1. Then I created the filesystem with mkfs.ext3 /dev/md0.
I then mounted md0 at /Video with mount /dev/md0 /Video/ All according to the guide.. Today I made a samba-share out of /Video/Rorschach to easily put files in there from my windows7-machine (the plan is to steam from my CentOS-server to my HTPC which hasn't arrived yet). I started to put movies in there. It went just fine for a while but then I got this message: [URL]... How is that even possible when df -h looks like this?:
I fail to install OpenSuse 11.3 with Raid-1. What I do: Partitioning -> Expert. sda1 type RAID do not format / mount, sda2 /(root) type RAID do not format / mount, sda3 /home type RAID do not format/mount. Clone disk to sdb. Raid -> add md126p1 type swap mount to swap, add->md126p2 type etx4 mount to /(root), md126p3 type ext4 mount to /home. Bootloader: GRUB, boot from MBR enabled, boot from / disabled. After installation the system does not boot and grub reports error that the specified filesystem can not be found.
I just installed Ubuntu server edition to my computer (brand new, no OS) and finished installation. In the terminal I used apt-get ubuntu-desktop to install a desktop interface.In my rig, I have two 500GB HDDs. I set them up through my computer BIOS as RAID1 drives, yet as I understand I still need to configure the Ubuntu software raid for it to work correctly. Unfortunately, I already partitioned my drives! I used the easy way (guided with LVM or whatever) and let it do it for me. Now, RAID1 is very important to me! Is there anyway to repartition the disks to use RAID1, or do I need to wipe my computer and reinstall Ubuntu?
I just experienced a HDD failure and while reorganizing the drives inthis machine I realized the benefits of UUID instead of /dev/sdX nomenclature. I am trying to determine the UUID of 2 disks that are assembled in a RAID1 array. right now they are /dev/sde & /dev/sdf with each only one partition. I tried ls -l /dev/disk/by-uuid but I get only the UUID of other disks, not the ones currently ID'd as sde & sdf. my mdadm.conf assembles several raid arrays all by UUID, but somehow, I cant recall how I got the UUIDs of the other HDDs at first...
I've got a couple of commercial NAS boxes and I'm wondering if they (ReadyNas duo, DLink DNS-323) or any other NAS is suitable for having their RAIDed disks moved to a software-based NAS. To be specific, I'm a big fan of the (largely) Debian-based Ubuntu. Can the aforementioned NAS drives be migrated to Ubuntu (e.g. using the mdadm Linux command)?
Secondly, is there any commercial NAS that can be migrated over? Incidentally, here is a link to somebody who succeeded in a migration:URL...My specific scenario I'd like to prepare for, is the eventual (sudden) death of one of the NAS motherboards.
I got a new computer (core i7, 6gb ram, 500gb hdd).I first installed Windows 7, because I remembered how Fedora 9 created a dual-boot screen nicely, while if I installed XP on a system with Fedora 9 it would just take over the boot and boot to XP.Anyway, so Windows 7 is installed, and I got the Fedora 11 DVD install.I partitioned the hdd as thus (on win7 installation): 93gb windows system, 300gb media partition, 90gb unformatted.When I boot the installation and select the "use free space" I get an error that there isn't enough free space on the disks to create the layout.I also tried to create a custom layout, but after a few errors I gave up on that what should I do to get Fedora 11 installed?
I want to buy another laptop. I want to use FC on it. I need bluray because I want to save data on bluray disks of 25 gb (better than dvds, more space). I have one option, a fujitsu that has bluray writer, big display (I have another one, other model, same size and it's working fine with Fc14, only some small issues).
Second option is a Dell (I had an older dell inspiron before, which worked fine with FC), more expensive, lacking a writer, but it can read bluray. If I take this, I would have to also buy an external bluray writer. Has nice specs, but I am really worried about the graphic card, whether I can make it work with Fc14. I saw some people having problems with linux and this graphic card, while searching on google.
Here's specs for both (sorry for the language from website, but I found in English other specs than these, I guess the specs can be read easily): Dell XPS L702x and FUJITSU Lifebook NH570 I am inclining more towards the Fujitsu, but haven't decided yet.
I'm trying to install Fedora 13 on my HP dv6 laptop and when I try to use the Use Free Space installation type, I receive a Partitioning Error: Could not allocate requested partitions - not enough free space on disks. Before I started the install, I used the Disk Management utility in Windows 7 to shrink the volume of the C: drive down to 242 GB and leaving 210 GB Unallocated. Here is what the screen looks like when I select Create Custom Layout (also receive the not enough free space error):
I have Dell Laptop 1545 which already installed windows 7 home premium which is also having two partitions one is reserved by the oem and another is for recovery partition and another 200gb i am using for windows 7 now i have left only 80gb hard disk. So I started to install the Fedora 12 in my laptop every thing is going fine but
at the time of creating the partitions iam unable to allocate the partitions the left 80gb i tried to select and tried for custom partition but to my surprise it is giving the following message "Could not allocate requested partitions: not enough free space on disks"
I am using CentOS 5.2. I am installing from disc on a machine with Intel Embedded Server Raid Technology. It has two 500 GB SATA drives. During the initial boot process, it sees that these two devices exist. However, after getting into the screen to partition and configure RAID, it just shows this:
Drive /dev/mapper/ddf1_MegaSR R! #0 (475879 MB) (Model: Linux device-mapper)
I want to do a RAID1 so that the disks are mirrored. However, I would expect to see both drives listed. I can select RAID to create RAID partitions, but I think I need to be able to see both drives in order to do this correctly.
I'm new to Centos and very new to RAID/hd setup. I have a old HP proliant G3 ML150. I have no drivers cd or other, only the server. I have created a RAID1 array (named SYSTEM) with 2 HD of 250GB from the controller and have installed Centos 5.2 (updated after to 5.6). The installation is ok. Now I have added 2 HD of 1TB each and have created another RAID1 array (named DATI) from the controller. This RAID is to store data files. (And next I have to add another RAID1 for backup, but this is to do next week). how can I format and add it to Centos so I can use it?
I have software RAID 1 on two physical discs. There are now 4 md -partitions (md0 ... md3), which are used in such as / and /home among others. Now current size of /home (md3) is starting to be full, and since / (md1) has more than plenty of free space I decided to fix the situation by shrinking / (md1) partition to free 40 Gigs of space and then growing /home (md3) partition for that 40 Gigs.
I already checked for some info using mdadm and got the following:
Now I would need some support on HOW exactly should I do this resizing since it is on RAID partitions.
Would it be good to use resize2fs to modify the filesystem sizes and mdadm to configure the partition sizes. Or could I perhaps get over this even easier by using GPartED (in case it supports my RAID)? Has anyone here done similar resizing on software RAID1 partitions?
The motherboard currently installed on my PC has a RAID Utility (Ctrl+I) at the startup that allow creating RAID1. But I already have a system installed with CentOS 5.4. In order to protect my data, I need RAID1. Can I add another Hard Drive now and have the data mirrored and synced onto both hard drives as if it was in RAID1 right from the beginning?
Not sure on what is going on here. The server is RAID1 through hardware RAID. It was running an unusual high load so I rebooted it. Now it won't boot up. I am getting these errors after the CentOS boot screen:sda: Current [descriptor]: sense key: Medium ErrorAdd.Sense: Address mark not found for data field
end_request: I/O error, dev sda, sector 3040555357 device-mapper: raid1: A read failure occurred on a mirror device. device-mapper: raid1: All sides of mirror have failed.
OS Version: CentOS 5.3 Motherboard: ASUS M3A78-CM (BIOS v.2003)
I have a single disk running the base OS and just installed 2 x Seagate 500GB SATA 3.0 drives in a RAID1 set that I would like to use for data storage. The OS sees the drives individually but not the RAID. Has anyone worked with a similar board and has any ideas what I need to do to get the OS to recognize the RAID1 array?
and I verify that /dev/md2 and /dev/md3 exist. But if I reboot the computer, these two devices are gone. I'm sure I am overlooking a step but I can't seem to find what it is. Could someone tell me what I need to do next?
I'm trying to load CentOS 5.5 on a new server with an Intel S5500BC motherboard using RAID 1. This board has a known problem with RHEL 5.x and the driver disk supplied has a fix. Here is the download for the driver [URL] Under the ESRT2_RHEL4-5_SLES9-1--11_v.13.21.2010_README file are directions in Section 3.1.3 on how to install the RHEL5x megasr driver. This works. The last thing replaces the ACHI driver with the megasr driver (paragraph 15) by loading the megasr.13.21.0614.2010-1-rhel50-u4-all.img in a temp file and then type "./replace_achi.sh". This step doesn't work and it is the critical one as it replaces the achi with megasr in the initrd image.
I installed a distro based on CentOS 5.5 (FreePBX distro FYI). It used an automated kickstart script to create an md RAID1 array of all the hard drives connected to the machine. Well, I installed from a thumb drive, which the script in interpreted as a hard drive and thus included in the array. So, I ended up with three md arrays (boot, swap, data) that included the thumb drive. Even better, it used the thumb drive for grub boot so I couldn't start up without it. I was able to mark the USB drive as 'failed' and remove from each array, and even change grub around to boot without the usb drive, but now each of the arrays is marked as degraded:
I've got a mailserver set up in a raid1 array.I shut down the system to install a CD-ROM drive but forgot to change the master/slave settings (I know, don'tt say anything) and didn't realize it before Centos started booting up, so it booted the hdc drive from the array.I rebuilt the array using mdadm without any apparant issues but on subsequent bootup, I get the following error :
There doesn't seem to be any side effects to this but since that didn't happen prior, I figure there's probably something I didn't do properly since I'm fairly new to the linux world.My raid array was originally set up by the Centos instal software and is set up like this :
hda1 + hdc1 = md0 (boot) Hda3 + hdc3 = md2 (/)
The other partitions are of the same size on each drive and are swap partitions.
PS : The drive is SMART capable and no errors appear during a self-test.
edit : Clonezilla also fails to boot properly although I don't know if its due to a software raid array in the first place or the errors in the filesystem. When only one drive was detected because of the jumpers, it booted properly.
I have installed a 2TB drive in my dual PIII 866 with 750MB ram. The drive is properly installed and I have configured the drive with 1 partition in RAID1. The array loads fine, but when I add the entry to mount the /dev/md2 /data/repository the following error occurs The filesystem size according to the superblock is 488378000 blocks The physical size of the device is 488377986 blocks Either the superblock or partition table is likely corrupt I have run fsck manually with no errors reported. I have removed the partition and rebuilt the array. The array assembles properly and I can manually mount the /dev/md2, but as soon as I add the entry to the fstab I get dropped to a shell after a reboot. Not sure where to go now?
I have 2 WD20EARS hard drives on the way (2 TB green WD disks with 4k sectors) and I'll be installing Centos 5.5 in RAID1 on them (2 partitions, one 16 GB / at the beginning and the rest in its own partition). I read the following thread: [URL]
and it seems that I might be having problems with the 4k sectors (Advanced Drive Format in WD lingo). I'm confused as to what exactly to do. I was thinking of downloading Fedora 14 Live CD and partitioning there and then switching to Centos 5.5 to install. Will that work? Seems I want the md 0.9 metadata because it doesn't have the space limit for me (2 TB) and it's stored at the end of the partition so it avoids alignment issues. Will I be able to make that happen with Fedora 14?
I've faced the problem with server freeze on heavy write.
CentOS 5.5 x64_86 with latest updates and kernel (2.6.18-194.32.1). Also tried 2.6.18-194.26.1 and 2.6.37-2 from ELRepo with the same results. CPU: Intel(R) Core(TM) i7 CPU 920 @ 2.67GHz Memory: 3 x 2Gb DDR3. HDDs: 2 x Western Digital WDC WD1002FBYS-02A6B0
I'm a Linux newbie but familiar with computers in general. I can install 9.4 64 bit (but no network to update from), but 10.4 and 10.10 both fail -- I cannot get the Allocate drive space screen to list the available drives. Just a blank panel. The drive is listed in the Boot loader panel. The LiveCD works. Disk utility and Gparted are both available. At one point I even managed to mount the 9.4 file system on the LiveCD, I think using "sudo mount"
Created abt 16.4 gb free space using Disk Management in vista. I read the sticky on installing from the live cd and did accordingly. Whatever Partition i create first (boot or /), it gets done. However, when tryin to create the 2nd partition, i get "Could not allocate requested partitions:Not enough free space on disks." Cant proceed any further.
Yesterday I installed a new server with a large partition for my XEN images. This partition is a about 930GB. The installation tooks ages and after he finished I was finding out why that is. The SoftRAID1 I configured is rebuilding the large partition.
We have had a hardisk crash in our RAID1 webhosting server running CentOS5 and Plesk. We first realized something was wrong when our main site didn't load but showed MySQL errors. We then found out that the system was in read-only state. Something that also happened the day before yesterday, but we could fix with a FSCK. Then the system worked well til around 18 hours later when it crashed with the same sympoms. So, we rebooted the server and wanted to do a filesystem check again. But the HDD wouldnt even load. It was gone. Unfortunatelly it was not realized that the second disk in the system was also not working any more for some time now. Fortunatelly we had our main site backed up externally though. So we could re-install a fresh box and mounted the two drives to the system. We checked the harddisk. One is practically empty (the older one), the other has almost only files in 'lost + found' but these are all "numbered", no real filenames or so.
System Specs 3.2 GHz Dell GX280 Sans Digital TR5M-B in JBOD configuration (PCIe 1x) CentOS 5.4 installed via QmailToaster instructions. Up to date via YUM. No GUI (no KDE/Gnome), command line only.
My old server died and reinstalled based on above with no problems. All is running well. Box is used as mail/web/home disk network. I want to be able to spin down the drives in the Sans Digital box when idle as they are used mostly for bulk storage (video editing, etc...). The external enclosure on my old server automatically spun down idle drives so this has never been a concern. Google has not helped me so far (operator error?). I did see "hdparm" may work if the drives support it but I am not sure if this is the nest solution.
I've downloaded 7 ISO images in order to install CentOS as a Virtual Machine on Fusion. First disk goes through successfully, but when I try to "insert" 2nd disk, I get an error message "unable to access the cdrom". I've ejected 1st disk, mounted second. I tried all possible scenarios I could think of, but still getting an error message.
I see the disk mounted in a terminal: /dev/disk1 600600 600600 0 100% /Volumes/CentOS_5.5_Final