Server :: Disks Not Mounted After Unexpected Power Down
Apr 30, 2011
For some reason we have some network storage not mounted properly after an unexpected restart (ie OS or memory issue) and the subsequent reboot. Has anyone come across this before? It's a bit strange that certain disks are not mounted.Anyway, to detect this, can we insert an extra line into the /etc/inittab to perform post start checks. For example, I have inserted an extra line before the runlevel scripts in /etc/inittab file.
The line will call /usr/local/fstab-check.sh to perform a check on the disks mounted via /etc/fstab. Am thinking of comparing the output from the command "df" versus the content of /etc/fstab.
A small combined webserver and fileserver with focus on power consumption and storage size, what would be a cheap and stable solution to this in terms of hardware ? I'm not interested in NAS solution only servers ..
My university has a system set up where each class has its own directory for the professors to publish and receive files. These directories reside under a directory in /home which happens to be a mounted disk and are accessible anywhere in the system by referencing ~csXXX where csXXX is a sub-directory within that home directory. How are they doing this and how can I do it on my own systems?
I have a 3ware controller that has a RAID 1 of two SATA disks.After an outage, the linux box (which is running ubuntu), restarted and the partition is now mounted read only. I only have the "/" mount point (this is a test server).Now, if I go to the 3ware controller by pressing ALT-3 while booting, I don't see any indication that there is something wrong with the disks.If I let the computer boot, I'm asked by fdisk if I want to fix/ignore/etc the inconsistencies found.
I am trying to build a script that runs on AIX, Solaris, and various flavors of Linux as well. It is very simple, or so I thought. I have the AIX and Solaris servers working well. Next I tried to get my workstation to run the script and it bombed. I am running Fedora 12 x86_64, but that isn't really important. I have various flavors of Linux. Some CentOS, RH, and even a Gentoo server. The script I have runs every hour on all servers. I can push out changes and install software with it, but the Linux WS won't run it.
Here is my script:
And here are the errors:
Is this just an issue with going between ksh and bash?
In my case, it occurs intermittently when trying to connect using putty's command line client 'plink.exe' to a linux machine using ssh. For some testing purposes, a linux server is setup with 8K virtual ip addresses and the testing scripts will connect to these ip addresses through ssh using plink.exe. There will be concurrently 25 ssh sessions launched to the server. While it works for sometimes, it throws 'server unexpectedly closed network connection' error intermittently. If the same ip address is tried again manually, it works. While the ssh server is accepting connections, this error is thrown sometimes and unable to find out the reason from the logs. It is not that the ssh connection is completely blocked. It gives this error message while connecting to some of the virtual ip addresses but works some times.
The hosts.deny, sshd_config are tuned to allow the connections, allow more sessions etc.
I have servers which contain SATA disks and SAS disks. I was testing the speed of writing on these servers and I recognized that SAS 10.000 disks much more slowly than the SATA 7200. What do you think about this slowness? What are the reasons of this slowness?
I am giving the below rates (values) which I took from my test (from my comparisons between SAS 10.000 and SATA 7200);
dd if=/dev/zero of=bigfile.txt bs=1024 count=1000000 when this comment was run in SAS disk server, I took this output(10.000 rpm)
(a new server,2 CPU 8 core and 8 gb ram)
1000000+0 records in 1000000+0 records out 1024000000 bytes (1.0 GB) copied, 12.9662 s, 79.0 MB/s (I have not used this server yet) (hw raid1)
I have/had a PC with several hard drives, and a mix of ubuntu and windows on multi boot.The old boot drive died screaming, and I need to start again. (But my data is safe! yay!)
Is there anything special about which drive can be the main drive to start booting from? Or to put it another way, can I install to any of the other 3 and expect it to work, or do I need to switch them around so a different drive is on the connections for the recently dead one?
I am very new to linux, and I have a question regarding the filesystem check (fsck). The power recently went out and when I tried to restart linux the following error appears:
*/dev/sda1 contains file system w/errors, check forced it then goes on to say..
*An error occured during the file system check. Dropping you to a shell; the system will reboot when you leave the shell. Give root password for maintenance (or type Control-D to continue) I wasn't sure what to do, but checked some other online forums and they suggested running fsck manually - so I typed in the root password - and used the command, "fsck -A -V ; echo == $? ==" it then gave the following message
*WARNING!!! Running e2fsck on a mounted filesystem may cause SEVERE filesystem damage *Would you like to continue (y/n)
Again, I wasn't sure what to do so i just checked no. I then manually turned off the computer and was prompted at the beginning to press Alt-3. I was brought to another screen and it informed me one of the drives was degraded and suggested rebuilding the array. I tried doing this, but it still brings me back to the original error of, "/dev/sda1 contains file system w/errors, check forced," and the process continues.
Also, when I tried to rebuild the array, I didn't backup any of the data on our home directory before doing this (which was probably a big mistake). After being prompted to type the root password, I was able to give the ls command and look at all the directories...the home directory where our data was stored was empty and I am afraid I may have lost some information. Is there a possibility that data was lost when I was trying to rebuild using the old drives?
Perty self explanitory, I'm trying to boot up ubunto server edition 10.10 from a USB stick (pen drive) and it works fine until step 3 when it says "Your installation CD-ROM couldn't be mounted. This probably means that the CD-ROM was not put in the drive. If so you can insert it and try again
Try again to mount the CD-ROM? <Yes> <No>"
I don't have an installation cd-rom! I put the .iso file onto a memory stick, inserted it into my old computer (which i will turn into a server later) , restarted it and got past the first two slides and got this error? Do I have to put the iso on a cd-rom?
I have a web application(java 5, mysql server) installed on Tomcat on Red Hat 5.Part of the functionality is to allow the upload of files to a windows share. I have mounted the share using theAny file I create here from the terminal gets sions."-rwxrwxrwx"The problem is that when the application server creates the file it gets created with the following."-rw-r--r--"How can I get files created from the server to be "-rwxrwxrwx"
I have built a couple RAID's, but I'm uncertain of how I should format the partitions of the raid. Should I format partitions on each disk, and then add them to a raid, or should I create a raid on unformated disks and then format the raid as a partition? Does it matter, and are there performance/reliability issues? I'm creating a RAID-5 using 3 SATA disks on RHEL for user data area.
We use powerpath on RHEL servers, with at least 2 paths to each disk, as you know the powerpath creates a pseudo devices called emcpowerX. I was wondering if there is any overhead if we don't use the lvm.conf to filter out all sdX disks? Except the "duplicate PV found" messages on LVM, does it delays the boot of the server?
I am connecting a complete set of four hard disks externally to my linux server.I have certain queries on that
1. How to make it detected. Will the cd which had come with that will install the required drivers? 2. How the cd drive will work out in linux? 3. How to configure all the 4 HDD into one LVM? 4 How to remove an already installed software from a linux system?
I've installed, overwrote, and deleted the partion with US to install again.Each time I could'nt login.I used 2 versions, 9.10 + 8.10. It can't be the login info, but the problem is probably that I hadn't installed any packages, e.g. Mail or Fileserver.Could the HDD not be mounted? I'll attach a GParted screenshot.
I have to move all the files and directories between 2 file systems. Is it good practice to move them at once or first copy them and then move ? How to do this to preserve the permissions and directory structure ?
I have been running a server with an increasingly large md array and always been plagued with intermittent disk faults. For a long time, I've attributed those to either temperature or power glitches. I had just embarked on a quest to a) lower case and drive temperature. They were running between 43 and 47C, sometimes peaking at 52C, so I've added more case fan power and made sure the drive cage was in the flow (it has it's own fan, too). Also, I've upgraded my power supply and made very sure that all the connectors are good. The array currently is a RAID6 with 5 Seagate 1,5TB drives.
When everything seemed to be working fine, I looked at my SMART logs and found that two of my drives (both well over 14000 operating hours) were showing uncorrectible bad blocks. Since it's RAID6, I figured, I couldn't do much harm, ran a badblocks test on it, zeroed the blocks that were reported bad, figuring the drive defect management would remap them to a good part of the disk and zeroed the superblock. I then added it back to the pack and the resync started. At around 50%, a second drive decided to go and shortly thereafter a third. Now, with two out of five drives, RAID6 will fail. Fine. At least, no data will be written to it anymore, however, now I cannot reassemble the array anymore.
Whenever I try I get this: Code: mdadm --assemble --scan mdadm: /dev/md1 assembled from 2 drives and 2 spares - not enough to start the array
Which is not fine. I'm sure that three devices are fine (normally, a failed device would just rejoin the array, skipping most of the resync by way of the bitmap) so I should be able to reassemble the array with the two good ones and the one that failed last, then add the one that failed during the resync and finally re-add the original offender. However, I have no idea how to get them out of the "(S)" state.
At times I see heavy hard-disk activity slowing down system.I have enough RAM and enough of it free.Apparently I am not doing anything that should cause this.'vmstat' shows high numbers for "bi bo" I want to find out which processes are doing "bi bo".How can I zero in on the particular processes?What command?
Is SUBJ possible? For example: I have 1 disk with ext2 (ext3) partition and information written on it. I want to make a mirror using the disk. I put this one and another unformatted one into a server, create an appropriate /etc/raiddraive, and do mkraid. Shall I get a mirror witn intact information from the first disk? If not, are there any ways to do so?
I'm trying to mount a mounted NFS share. I have a server that connects through VPN to a network, that has the NFS share exported. I am able to successfully mount the NFS share on that server, in /media/iSCSI. Now I want to share this NFS share with other servers, that are on the same network as the VPN-ed server, but are not connected to VPN. When I try to export the mounted share, I get:
I have a DHCP/PXE server behind a firewall. It mounts partitions on the file server on the corp. network on the other side of the firewall. Every box that PXEs also mounts partitions on the main file server.
I was hoping I could change them to mount from the DHCP/PXE server, so that server could cache and cut down on the requests through the firewall, as well as the sessions that the firewall must track.But it seams a little strange to try to export directories that are simply NFS mounts on another server already.
I have a server running RHEL6 and a virtual machine also running RHEL6. I created a directory /home/data on the server and another on the VM. When I mount the host directory on the VM, I am not able to change the ownership/permissions through the VM no matter what. The ownership is set to "nobody" and I can't even change it to root.
I have a chrooted account setup for my ssh server. However, I am trying to allow this user read only access to access on a mounted hard-drive and more specifically a specific folder from that mounted drive. I would also like to have this drive be mounted for me in my normal environment with write access.
I have a low maybe medium experience with linux and no experience with unix and migrating from freenas to Ubuntu.I'm trying to transfer 12 TB of data from UFS file system to ext3 file system under Ubuntu 10.04. The server has two 16TB virtual disks (raid10) and one 4TB virtual disk (raid10 too).I mount the UFS virtual disk in Ubuntu with this command:
Code: mount -r -t ufs -o ufstype=ufs2 /dev/sda1 /home/ufsdisk I am able to read only 6TB of data instead of 12TB. The missing folders can be seen in