General :: RAID Disk Inactive After Reboot

Aug 12, 2010

I just configured two raid setups but after a reboot they are not mounted and seem to be inactive.

md127 = sde1, sdf1 and sdi1 (raid 5)
md0 = sda1 and sdh1 (raid 0)
root@server /]# cat /proc/mdstat
Personalities :
md127 : inactive sdf1[1](S) sde1[2](S)
78156032 blocks
md0 : inactive sda1[0](S)
488382977 blocks super 1.2
unused devices: <none>

[root@server /]# fdisk -l | grep "Disk /"
Disk /dev/sda: 500.1 GB, 500107862016 bytes
Disk /dev/sdb: 80.0 GB, 80026361856 bytes
Disk /dev/sdc: 122.9 GB, 122942324736 bytes
Disk /dev/sdd: 160.0 GB, 160041885696 bytes
Disk /dev/sde: 40.0 GB, 40020664320 bytes
Disk /dev/sdf: 40.0 GB, 40020664320 bytes
Disk /dev/sdg: 1500.3 GB, 1500301910016 bytes
Disk /dev/sdh: 251.0 GB, 251000193024 bytes
Disk /dev/sdi: 40.0 GB, 40020664320 bytes
Disk /dev/sdj: 500.1 GB, 500107862016 bytes

[root@server /]# cat /etc/mdadm.conf
DEVICE /dev/sdi1 /dev/sdf1 /dev/sde1 /dev/sda1 /dev/sdh1
ARRAY /dev/md127 UUID=5dc0cf7a:8c715104:04894333:532a878b auto=yes
ARRAY /dev/md0 UUID=65c49170:733df717:435e470b:3334ee94 auto=yes

As you can see they now show up as inactive. And for some reason sdi1 and sdh1 are not even listed. What can I do to get them back? To make matters worse I placed some important data on them, and even if I was clever enough to keep an extra copy on another drive, guess which drive that was? So, I need to get them activated as is (at least so I can get the data of them) before I can rebuild them from scratch. I'm running Mandriva 2010.1 and rated tehm using the built in disk partitioner.

View 14 Replies


Ubuntu :: Raid Array Inactive After Reboot?

Apr 24, 2011

I had some issues with my RAID6 array (with 15 disks), where 5 disks got disconnected (each five disks is connected to the motherboard via 1 SATA cable), which brought down the RAID array. I fixed this problem via readding the disks and running the array:


mdadm -R /dev/md1

However, after rebooting, the array appears inactive, and I have to go through the same motions to fix this and make it active.
The array is present in the /etc/mdadm/mdadm.conf, though also 2 other raid arrays (3 arrays total):


DEVICE partitions
# auto-create devices with Debian standard permissions
CREATE owner=root group=disk mode=0660 auto=yes


View 1 Replies View Related

General :: Get An Inactive RAID Device Working?

Mar 9, 2010

After booting, my RAID1 device (/dev/md_d0 *) sometimes goes in some funny state and I cannot mount it.

* Originally I created /dev/md0 but it has somehow changed itself into /dev/md_d0.
# mount /opt
mount: wrong fs type, bad option, bad superblock on /dev/md_d0,


In /proc/partitions the last entry is md_d0 at least now, after reboot, when the device happens to be active again. (I'm not sure if it would be the same when it's inactive.)

Resolution: as Jimmy Hedman suggested, I took the output of mdadm --examine --scan:
ARRAY /dev/md0 level=raid1 num-devices=2 UUID=de8fbd92[...]

and added it in /etc/mdadm/mdadm.conf, which seems to have fixed the main problem. After changing /etc/fstab to use /dev/md0 again (instead of /dev/md_d0), the RAID device also gets automatically mounted!

View 5 Replies View Related

Fedora Networking :: Eth0 Says Inactive On F10 Reboot

Apr 1, 2009

I just installed Fedora 10. I entered my static network info in Network Configuration. When rebooting, the eth0 says Inactive. I've read many articles about this problem on earlier versions and tried many of the suggestion, but have not found a definite way of solving this. Is there a better way or a workaround to get this machine on the network?

View 6 Replies View Related

Slackware :: RAID 5 Boot - LVM Stays Inactive

Nov 5, 2010

I'm somewhat stuck I fiddle with intel ss4200 NAS where I managed to install Slack 13.1 on some spare IDE HDD that was lying around (instead of that crappy 256 MB DOM). Anyway, everything works except one thing...

The setup is: 1x IDE HDD with Slackware on it. 4 x 1TB drives in RAID 5, which when mounted make 3 different logical drives. Everything works, /dev/md0 is created, pvc on top of it with 3 different drives. Well at least until I reboot. After reboot /dev/md0 is not automatically assembled and due to that - lvm stays inactive. Of course, I can make a script and put it in rc.local that will activate and mount what I want but I'm sure that there is one more elegant solution for that.

At the moment I need to issue: mdadm --assemble /dev/md0, vgchange -ay, mount -a and vgchange -an at shutdown. I checked parts concerning LVM in rc.S but I'm clueless. Kernel on the system is 2.6.36, mdadm - v2.6.9 and LVM version: 2.02.64

View 2 Replies View Related

General :: RAID 1 With 2 Disks Starts Degraded After Reboot From 3rd

Jun 20, 2010

Basically, I installed Debian Lenny creating two RAID 1 devices on two 1 TB disks during installation. /dev/md0 for swap and /dev/md1 for "/"
I did not pay much attention, but it seemed to work fine at start - both raid devices were up early during boot, I think. After that I upgraded the system into testing which involved at least upgrading GRUB to 1.97 and compiling & installing a new 2.6.34 kernel ( udev refused to upgrade with old kernel ) Last part was a bit messy, but in the end I have it working.

Let me describe my HDDs setup: when I do "sudo fdisk -l" it gives me sda1,sda2 raid partitions on sda, sdb1,sdb2 raid partitions on sdb which are my two 1 TB drives and sdc1, sdc2, sdc5 for my 3rd 160GB drive I actually boot from ( I mean GRUB is installed there, and its chosen as boot device in BIOS ). The problem is that raid starts degraded every time ( starts with 1 out of 2 devices ). When doing " cat /proc/mdstat " I get "U_" statuses and 2nd devices is "removed" on both md devices.

I can successfully run partx -a sdb, which gives me sdb1 and sdb2 and then I readd those to raid devices using " sudo mdadm --add /dev/md0 /dev/sda1 ". After I read devices it syncs the disks and after about 3 hours I see fine status in mdstat. However when I reboot, it again starts with degraded array. I get a feeling that after I read the disk and sync array I need to update some configuration somewhere, I tried to " sudo mdadm --examine --scan " but its output is no different from my current /etc/mdadm/mdadm.conf even after I readd the disks and sync.

View 1 Replies View Related

General :: Hard Drive Dropping Out Of RAID On Reboot

May 30, 2011

Every time I reboot my server, one of my hard drives drops out of the RAID5 array. I'm pretty sure that there's nothing wrong with the drive itself. I bought all three drives at the same time, and they are identical in make/model/capacity. While the server is running, it's smooth sailing. However, whenever I shut down or reboot, I get an email message that the array is degraded. It's always /dev/sda1 that drops out of the array. I can always rebuild the array by adding the partition back in, but it's a bit of a pain. Any suggestions on how to troubleshoot this?

View 13 Replies View Related

General :: Unable To Access Intel Fake RAID 1 Array In Fedora 14 After Reboot

Jan 6, 2011

1st I am relatively new to linux (but not to *nix). I have 4 disks assembled in the following intel ahci bios fake raid arrays:

2x320GB RAID1 - used for operating systems md126
2x1TB RAID1 - used for data md125

I have used the raid of size 320GB to install my operating system and the second raid I didn't even select during the installation of Fedora 14. After successful partitioning and installation of Fedora, I tried to make the second array available, it was possible to make it visible in linux with mdadm --assembe --scan , after that I created one maximum size partition and 1 maximum size ext4 filesystem in it. Mounted, and used it. After restart - a few I/O errors during boot regarding md125 + inability to mount the filesystem on it and dropped into repair shell. I commented the filesystem in fstab and it booted. To my surprise, the array was marked as "auto read only":


and the partition in it was not available as device special file in /dev:


View 1 Replies View Related

General :: Raid - RAID1 With Only One Disk

Oct 27, 2010

I have a hypothetical situation in which I installed my operating system using a RAID1 mirror. At some point I decided that this setup was overkill, my machine isn't system critical, I value doubling my storage space more than speedy recovery, I'm doing routine backups, etc...

Short of backing up my system volume and repartitioning, or otherwise starting over, is there a way I can reconfigure my RAID1 array to only expect one disk so that mdadm no longer reports a Degraded state?

View 3 Replies View Related

General :: Setup RAID 5 And One Spare Disk

Aug 4, 2010

I want to build a 6xSATA RAID 5 system with on of the disks as spare disk. I think this give me a chance of 2 of 6 disks failing without losing data. I am right?
Hardware: Intel ICH10R
First I will creat a 3xSATA RAID 5, after I will add the spare disk and after that I will add the others disks. This is what I think I should do.

Step 1:
Create RAID Device
mdadm --create --verbose /dev/md0 --metadata 1.2 --level=5 --raid-devices=3 /dev/sda1 /dev/sdb1 /dev/sdc1
I read that "--metadata 1.2" is the best option. It is true?
Create filesystem on the RAID device

Using this method of calculation:
* chunk size = 128kB (for RAID 5)
* block size = 4kB (recommended for large files, and most of time)
* stride = chunk / block = 128kB / 4k = 32kB
* stripe-width = stride * ( (n disks in raid5) - 1 ) = 32kB * ( (5)- 1 ) = 32kB * 4 = 128kb
mkfs.ext3 -v -m .1 -b 4096 -E stride=32,stripe-width=128 /dev/md0

Step 2:
Add spare-disk
mdadm --add /dev/md0 /dev/sdd1
Is this enough?

Step 3:
Adding disks:
mdadm --add /dev/md0 /dev/sde1
mdadm --grow /dev/md0 --raid-devices=4
fsck.ext3 /dev/md0
resize2fs /dev/md0

View 1 Replies View Related

General :: Where Is The Missing Disk Space On Software Raid

Nov 3, 2010

Purchased (4) 2TiB Drives (actual disk space) and created a RAID5 array expecting to have 6TB of useable disk space, however actual useable space is 5.46TiB.

So, the question is where did the disk space go?

First off, I can say for certainty the disks actual useable is verified at 2TB each have mounted and formated on a non-linux system (OSX).

Disks - 2TB Per disk, Tested HFS, Actual 2TB Useable
root@server:/server# fdisk -l 2>/dev/null | egrep "sd[hijk]" | grep Disk
Disk /dev/sdh: 2000.4 GB, 2000398934016 bytes
Disk /dev/sdj: 2000.4 GB, 2000398934016 bytes
Disk /dev/sdk: 2000.4 GB, 2000398934016 bytes


View 2 Replies View Related

General :: Debian Software RAID 1- Boot From Both Disk

Mar 15, 2011

I newly installed debian squeeze with software raid. The way I did was, as also given in this thread.

- I have 2 HDD with 500 GB each. For each of them, I created 3 partitions (/boot, / and swap)
- I selected the hard drive and created a new partition table
- I created a new partition that was 1GB. I then specified to use the partition as a Physical Volume for RAID. and used for /boot and enabled bootable.
- Created another partition, which is of 480 GB, and then specified to use the partition as a Physical Volume for RAID. and used for /.
- Created another partion and used for swap

Then RAID configuration:
Through Configure RAID menu -> create MD device ->
(2 for the number of drives, 0 for spare devices)
Next select the partitions you want to be members of /dev/MD0. I selected /dev/sda1 and /dev/sdb1 (for /boot)
Next select the partitions you want to be members of /dev/MD1. I selected /dev/sda6 and /dev/sdb6 (for /)
And no RAID for swap partitions

'Finish Partitioning and write changes to disk' --> Finish the rest of the install like normal. Everything is ok now, except I am not sure how to test my raid config. When I pull the power of the HDD, it only boots from one disk. I read in some forum that I may have to install GRUB manually on the other. In Debian Squeeze, there is no grub command. Not sure how to make my software raid bootable from both disk. I configured /boot partitions of both disks to be boot=yes. Not sure whether that is ok.

View 2 Replies View Related

General :: RAID Disk Failure Interrupt Notification?

Apr 4, 2010

I have installed a Fedora Core 12 Linux system onto a RAID 1 file system. I now need a way of getting an notification if the disk fails. Is there an SNMP MIB that covers Intel RAID? I have done the searching but still the answer alludes me.

View 1 Replies View Related

General :: No Automatic Rebuild Of RAID 5 After Replacing Bad Disk

Dec 12, 2009

I have a 5 disk raid 5 array that is composed of SATA A:0,1; SATA B: 0,1, and SATA C:0, and one of the disks (SATA A:0) recently went bad on me. I have an ICP raid controller that is about 5 years old. I replaced SATA A:0. After rebooting, I went into the controller and verified that it saw the disk in the hard-disk info section...there I noticed that in the "status" section, that the SATA C:0, SATA B:1 disks were listed as being "in array", the SATA A disks were blank, and the SATA B:0 disk was listed as "fragment". When I go into the "repair array" section, the controller tells me that there are no arrays that are in failure, error, or need to be rebuilt.

This puzzles me, as I thought the controller would know that the array needs to be rebuilt after replacing the disk and I don't see a way to initiate a rebuild. If I just let the server boot after replacing the disk, then I get back that there are the correct number of disks in the raid 5 and that it is ready, however, the screen then goes blank and I get a blinking cursor and the system seems to hang. There are no activity lights on any of the drives associated with the raid 5, which makes me think that the system is not rebuilding the array at this point.

View 3 Replies View Related

General :: MDADM Error: Creating A RAID 1 RAM Disk?

Apr 7, 2011

I am trying to create a Raid 1 ram disk. Below are the commands I used:

[root@abidbodal dev]# mke2fs -m 0 /dev/ram8
[root@abidbodal dev]# mount /dev/ram8 /mnt/rd8
[root@abidbodal dev]# mke2fs -m 0 /dev/ram9


View 3 Replies View Related

Hardware :: Fedora 11 RAID 1 - Disk Failure - Boot From The Single Working Disk?

Oct 16, 2009

my Fedora 11 system is not starting anylonger. It stops with the message:


VFS: Can't find ext4 filesystem on dev dm-0

The system told me since a while, that a lot of the sectors of one disk of the (software) RAID compound are failed already. So tried to disconnect each of the disks and start them separately. Unfortunaltly this is not working (for one its is not working at all, the other wents the same far as with both), when I tried to recover the system with the Fedora DVD, it said no distribution found. I am quite new and do not know so much about linux system, so i do not know what further information you could need. Maybe it can be important, that both disks are encryped (the system wents so far, that I can type in the password).

View 2 Replies View Related

Ubuntu Servers :: HW RAID Disk Shows Up In Fstab But Not In /dev/disk/by-uuid?

Jun 28, 2010

I have an SiI hardware SATA RAID card, with two 500GB disks in mirrored RAID configuration. When I first plugged them in and set it up, things seemed to work ok, but on boot the raid controller told me that the RAID needed rebuilding, and it would happen automatically after POST. So I didn't worry about it, and the drive mounted fine, and it's been that way for years. I just went in and manually on-line rebuilt the RAID in the controller's BIOS, and now when I boot into Ubuntu, both disks show up in fdisk, but neither show up in /dev/disk/by-uuid. Am I missing something?

View 9 Replies View Related

General :: Mount A Single RAID 1 Disk / Partition As Ext3?

Jul 7, 2011

I need to copy data from a single HD, which used to be part of a Linux RAID 1. I've googled around, but can't find any clue how to mount partitions from this single HD.

Background: The HD comes from a linux based NAS box Synology DS207+. The NAS uses ext3 as filesystem. Both NAS disks are fine, but the other NAS hardware is dead and not worth repairing or replacing.

View 1 Replies View Related

Ubuntu Installation :: Migrate Working Single Disk System To Existing RAID Array Using Disk UUIDs

Aug 1, 2010

I had done a new lucid install to a 1 TB RAID 1 array using the alternate CD a few weeks back. I messed up that system trying to some hardware working that lucid doesn't have drivers for yet, so I gave up on it and reinstalled to a single 80 GB disk that I now want to move over to the RAID array.

I moved all of the existing files on the array to a single folder, then copied all of the folders from the 80 GB disk over to the array with permissions and symlinks (minus the contents of /proc and /sys, which I created empty).

These are the commands I used:


p -a -d -R -v -t /media/raid_array /b*
cp -a -d -R -v -t /media/raid_array /d*
cp -a -d -R -v -t /media/raid_array /e*
cp -a -d -R -v -t /media/raid_array /h*


I tried to change fstab to use the 689a... for root, but when I try to boot, it's still trying to open /dev/disk/by-uuid/412d...

So then I booted from the single disk again and chrooted into the array, then ran update-initramfs -u. I got 3 "grep: /proc/modules: No such file or directory" errors, and "cat: /proc/cmdline: No such file or directory"- so I created directory /proc/modules, created an empty file /proc/cmdline, and ran the initramfs update again. Then I tried to shut down, which hung (probably because I was doing all of this from a terminal window in Gnome), so I killed the power after a couple of minutes.

It's still trying to use /dev/disk/by-uuid/412d... to boot.

What am I missing? I assume I just have to change the UUID to mount as root, but I don't know how.

View 2 Replies View Related

General :: Safe To Tune2fs Mounted Volume To Prevent Disk Checks At Reboot?

May 12, 2010

I have some large volumes that I don't want to automatically be e2fsck'd when I reboot the server. Is it safe to change maximum mount count to -1 and check interval to 0 while a volume is mounted, or will that cause problems to the file system?

View 2 Replies View Related

General :: Scsi RAID Jbod And Arrays - Disk Utilization And The Corresponding Low Data Transfer

Jul 6, 2010

So I have a system that is about 6 years old running Redhat 7.2 that is supporting a very old app that cannot be replaced at the moment. The jbod has 7 Raid1 arrays in it, 6 of which are for database storage and another for the OS storage. We've recently run into some bad slowdowns and drive failures causing nearly a week in downtime. Apparently none of the people involved, including the so-called hardware experts could really shed any light on the matter. Out of curiosity I ran iostat one day for a while and saw numbers similar to below:


Some of these kinda weird me out, especially the disk utilization and the corresponding low data transfer. I'm not a disk IO expert so if there are any gurus out there willing to help explain what it is I'm seeing here. As a side note, the system is back up and running it just runs sluggish and neither the database folks nor the hardware guys can make heads or tails of it. Ive sent them the same graphs from iostat but so far no response.

View 1 Replies View Related

Ubuntu Installation :: Raid 0 Is Not Mountable After Reboot

Jan 7, 2010

I currently am setting up an htpc running Karmic. The problem I am having is getting my Raid 0 to be mountable. My raid is not my boot partition, but is for data storage. My setup is a zotac motherboard with three sata connectors. I have a 300 GB drive, my eSata port, and my DVD attached to these. In the PCIe expansion slot I have installed Syba 2 port Sata PCIe 1a card using the Sil3132 sata II host chipset. Off of this I have 2 1.5Tib Hdds that I am setting up as Raid 0. During the boot I enter the chipset BIOS and establish this as a Raid 0.

When I install Karmic it sees the Raid and my 300 GB drive so I install to the 300 GB Drive and everything works fine. I am able to boot to the Hdd and run the OS. I then installed GParted and setup a partition on my RAID as GPT since I want one large partition of 2.78Tib. I then Format it through GParted as ext4 and I am able to mount and access it. I then reboot the system, and can no longer mount the filesystem. What I found interesting is If I reopen GParted I can then mount it. I traced it down to the fact that the until I access GParted the Block Special Device (sil_bgabagabaedd1) does not appear in /dev/mapper. Everytime I reboot I need to go into GParted to restore the Block Special Device then it is mounted. I think I am missing something in the raid setup as to why it is not being retained. What have I missed? What do I need to do to retain the Block Special Device? Is there a boot config setting?

Edit: I did further research and found that if I do kpartx it will appear just as gparted, but on reboot vanishes. I found something similar at this thread but not comfortable in updating dmraid: [URL] I think it is related to gpt and I will try to use a smaller partition to see if the behavior changes.

View 1 Replies View Related

Hardware :: Mounting Raid Device After Reboot?

Mar 24, 2009

configure my raid devices so it can be started and mounted at start-up.

View 1 Replies View Related

CentOS 5 :: Raid 5 Grow Does Not Survive Reboot?

Jun 22, 2010

I installed CentOS 5.5. After install, I decided to put 3 identical disk for raid 5. All the disks are IDE disk. Then I put a sata disk and partitioned it to add another partition to the raid 5 array. Everything works fine until I rebooted my system. After reboot, the sata partition I added into raid 5 is showing removed. I had to readd it using "mdadm --add" to make raid 5 array works.

View 5 Replies View Related

Fedora :: Software Raid 5 - UUID And Md Number Changes On Reboot?

Jun 18, 2010

So I recently set up a fedora 13 server using software raid. Let me go over the initial install and maybe that will help explain why I'm running into problems with one of the arrays. During installation I had only 2 disks in the equipment (WD 750GB each) Partitioned them thusly:

Disk /dev/sda: 750.2 GB, 750156374016 bytes
255 heads, 63 sectors/track, 91201 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00062206


View 3 Replies View Related

Debian :: Raid - No Booting / Reboot The System Does Not Boot?

Nov 5, 2010

There seems to be a problem with Raid on Debian. I got a new Fujitsu Primergy TS 100 S1 server, with hardware Raid (and 2 disks) installed everything nicely over the net including GRUB - but when it comes to reboot the system does not boot.

Is there anybody here who knows about the problem?

View 1 Replies View Related

Ubuntu Servers :: RAID Volume Won't Activate After Reboot During Sync?

Jul 24, 2010

I've got a file server with two RAID volumes. The one in question is 6 1TB SATA drives in a RAID6 configuration. I recently thought one drive was becoming faulty, so I removed it from the set. After running some stress tests, I determined my underlying problem hadn't cleared up. I added the disk back, which started the resync. Later on, the underlying problem caused another lock up. After a hard-reboot, the array will not start. I'm out of ideas on how to kick this over.

$ cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]


View 4 Replies View Related

Fedora :: F14 - Very Annoyed With Software RAID Setup - Cannot Find Superblock At Reboot?

Nov 20, 2010

I recently upgrade to Fedora 14 from 13. It was an in-place upgrade. I can't recall for sure, but I do believe I had these problems in F13 before the upgrade. The F13 install was from a Live CD. Anyway, I have a three drive RAID 5 array setup - 3x 750GB. For some very annoying reason, each time I reboot my F14 system, it hangs with an error about not being able to find a superblock on /dev/md126 and /dev/md127. I have tried to stop and remove /dev/md126 and /dev/md127 but they always seem to come back. I have also noticed in the output of fdisk -l that drives sda and sdd like to swap places sometimes for an unknown (to me) reason. Any other output that is needed, please ask. I recreated the array just yesterday with:

mdadm --create --verbose /dev/md0 --level=5 --raid-devices=3 /dev/sdb1 /dev/sdc1 /dev/sdd1
I would cat mdadm.conf in /etc, but I removed it previously to try to figure out the problem and it was not


View 1 Replies View Related

CentOS 5 Hardware :: 20 TB Software Raid 5 Build - Locked Up And Array Would Not Restart After Reboot

Mar 25, 2010

I have a 10x2tb disk array that i'm trying to build into a single software raid 5 i have tried this 2 times now the first it made it to 58.7% and the machine locked up and the array would not restart after a reboot. On my 2nd try all was looking good until about 50% i noticed that the speed dropped in 1/2 and that ksoftirqd/2 is taking up a lot of cpu (about 90%) the md0_resync and md0_raid5 are also taking 60-90% when the build started they took 7%. when i do a dmesg i see a lot of the message compute_blocknr: map not correct.

For a little info on the physical setup this is running on an Atom 510 with 2GB of mem the drives are connected to an addonics 4-Port RAID 5 / JBOD SATA II PCI Controller using the sil3124 chipset. I'm using 2 addonics 5X1 SATA Port Multiplier connected to the controller to get the 10 drives attached. All drive show up and don't seem to have any issues. I'm running a fully updated as of 3/20/10 version of centos 5.4

I will let this continue to run over night but i expect it to be locked up by morning if it follows what the last attempt did.

View 9 Replies View Related

General :: Understanding Idle/Inactive Sessions And Whether Or Not Someone Is Logged In?

Jan 5, 2011

I'm trying to understand if anyone is left on a server. Basically I manage a simple linux server remotely used by 3-4 individuals. I can never tell if someone is actually on or not using w/who.What I'm seeing is some people having what appears to be inactive/lost VNC sessions. I don't understand idle fully, but I do believe a program (without user interaction) can clear idle back to 0, correct?Anyway, I'm asking because every now and then I need to reboot the server, and I do not want to interrupt any program working on calculations or waiting on having the data saved.

me@matrix:~> w
06:59:54 up 170 days, 9:13, 16 users, load average: 0.52, 0.16, 0.06


View 1 Replies View Related

Copyrights 2005-15, All rights reserved