I know that uptime prints the time a machine has been up and running, but is there an easier (reliable) way to get the date of the start up than counting down from this output?I tried looking around /proc, but didn't find anything of relevance. There's also a line like this on my dmesg: [ 0.673492] rtc_cmos rtc_cmos: setting system clock to 2011-03-14 14:26:52 UTC (1300112812), but I'm wondering if this method is distribution and kernel version agnostic.
I know several tools that allow tracking time spend on different tasks / projects.Is there any existing tool for very very simplified work-time-tracking.I am an employee, come to the office, switch on my laptop directly. I have mostly around 1 hr lunchtime, but sometimes less, sometimes more.At around 18:00 I want to type one command in the console (or simple GUI would also be okay of course) that tells me:"1 hour overworked. Go home now! (came at 8:00, 1 hour standard lunch-break)."
My computer has different time when booting to linux or Windows.How to make the time the same?My computer time is 10:57pm Apr 14 when booting to linux.My computer time is 2:57am Apr 14 when booting to Windows Vista Home Premimum SP2.Both OS are set to the same time zone (GMT-5. Eastern Time US & Canada).
We have 2 applications set as S96 and S98 at rc3.d and rc5.d simultaneously. Both applications create a system V shared memory segment by calling shmget.If the system boot at runlevel 5, both applications can obtain their shared memory segment id correctly, i.e. 98305 and 131074 individually. While there is a root owned segment id 32768 takes first seat on the list. This is the id list:
I am having dual boot system(windows 7 and Fedora 12).When i switch on my system.It show the the timer 3 sec in order to get boot selection window(means window which asks that what to start fedora 12 or windows 7).I want to increase this time from 3 to 10 sec.
I've been having a problem on my AMD based machine, 4cpu, gigabyte ga-ma78gm-s2h Mobo, 8GB mem, two 2 terabyte Sata HDs.One thing I've found is that any kernel after 2.6.32-17 has a randomness at boot time whether the system will completely boot or not.
For instance just today I downloaded and installed 2.6.32-24
It fails to boot (I've tried cold boot, warm boot).Running its repair also fails to completely boot.My experience is that if I keep trying it "may" eventually boot but I believe there was some change after 2.6.32-17-generic that's causing the problem.Because as with 126.96.36.199... which also fails to complete bootup many times... eventually my guess is that 188.8.131.52 will also boot "sometimes".But why does 184.108.40.206 always boot for me? Something changed and its not my setup.
I'm just wondering what the limits for time are. I have a program that always takes exactly 20 ms, so I assume this is the lowest it can measure, but I want to see if there's some sort of documentation of this.
I want to know if there is a way to know the uptime for a server after a reboot process, I need that information for a statistic, but I forgot to take the uptime before reboot the server, so I am looking for that information after the server is power on.
For some reason, every time I shut down and reboot my machine, the system time advances four hours. The local time here is EDT (-4hrs UTC). I have it set correctly in Yast and have the 'hardware clock is set to UTC' checked. However, for some reason, the next time I boot, the system time will be ahead four hours.
I'm running a dual-boot; Lucid and Win XP on a HP Pavillion.My time settings are about 8 hrs off between the two operating systems. If I correct the time in Linux, it will be wrong when I boot in Windows. If I correct it in Windows, it will be wrong again next time I boot into Linux.Besides the obvious solution of removing Windows from my machine (which I'm not ready for), what should I do to fix this?
How to make the system boot faster by removing the idle time between 5s to 10s? bootchart attached. It is Ubuntu10.04LTS by the way. One more hint, the screen black out for ~4s after "Begin: Running /scripts/init-bottom... Done." I don't know what is going on during that 4s, but my best guess is there is a way we can get rid of it. Bootchart can be found here:
LIVECD of 11.3 Gnome works fine in my old PC. I installed several times and each time system freezes/crashes during boot. I am linux newbie, so, I am trying to figure out my problems. In Livecd I noticed Hardware info list's my SATA disk as IDE. Driver Modules: "ata_piix"Attached to: #24 (IDE interface). Is it right? I have posted Hardware information output in 01: None 00.0: 10105 BIOS [Created at bios.186] - Suse 11.3 install problem My machine specs are: Intel DG31PR motherboard, Intel core2duo, 3 GB DDR2 Ram, 250 GB SATA harddisk
Average for the last 35 loops: 27236.66 cycles What could be causing the strange value of 3322259 us in method (a)? I could just use method (b) but I would like to know what is going on here... Btw, I am using a desktop - with Linux debian 2.6.26 #1 SMP.
I'm a little bit confused with partitioning the filesystem in Linux. the difference between creating the file system with fdisk and mkfs (when formatting the disk). I can't clearly tell my problem, so please look at this picture:
I just recently installed ubuntu 9.10 in my upstairs computer. It is a single boot system.Downstairs I have a dual boot system. I have windows vista and ubuntu 9.10 installed. It worked fine. I wanted to make this a single boot system and uninstall ubuntu 9.10. I cannot get rid of the grub bootloade