General :: Tail Command Does Not Refresh All The Time?
Oct 26, 2010
I am running a script with nohup and this generates a lot of logs.
In order to view the log I use tail -f nohup.out
The problem is that the info supplied by this command is not always the latest//sometimes I need to use the command again order to view the latest info added to the nohup.out file.
we are using thunder bird as email client to download all our emails from mail server.We normally use to click get mail tab in thunder bird for new messages.My question is is there a way to check from new emails through command line. is there any specific command to do this.
when im trying login in my web ftp with correct pass and username its automaticly refresh the page. I have instaled easy-wi for i can make servers, but ftp dont works. Second problem is cronjobs. i dont know how to enabkle for this panel.
How to refresh a page automatically?Say for example i need to refresh page in ubuntuforums to get new questions.I feel lazy to refresh the page often.Is it possible to refresh the page automatically in a specific time interval?I have tried ReloadEvery Firefox Add-on.But it refresh all the tabs.What i want is i want to refresh a page in a particular TAB.
I am trying to use tail -f and play a sound everytime a new line appears. I tried this: for i in tail -f myFile; do aplay alert.wav; done; Which kinda worked, the output is:
Playing WAVE 'alert.wav' : Signed 16 bit Little Endian, Rate 22050 Hz, Mono Playing WAVE 'alert.wav' : Signed 16 bit Little Endian, Rate 22050 Hz, Mono Playing WAVE 'alert.wav' : Signed 16 bit Little Endian, Rate 22050 Hz, Mono
But after 3 times it stops, and I would like to print the tail -f result and not the aplay result. How could I achieve that?
I was recently looking into using tail -f to monitor some text files like so: tail -f /var/sometext However, when I did some testing, it doesn't seem to work. What I did was I created a new file and ran: tail -f /home/name/text Then, I opened the log in vim and did some editing, saved it, and it seems that tail is not "seeing" the change.
The weird thing is, running echo "hello" >> /home/name/text seems to work fine (tail sees the change). I read somewhere this has something to do with file descriptors and new inodes being created when saving a file.
am facing a problem with tailing a log file. Logs of application located in one folder:applog_20100101_0200.log <--log until 2 am january 1applog_20100101_0456.log <--log until 4:56 amapplog.log <-- current logApplication can change log when ever it wants to. I need to monitor this log, what i do:tail -f applog.logBut when app changes log my tail just stops. How can i tail applog.log all the time with out stops?
I'm timing how long it takes to run a command foo. I'm looking to append the results from the time command to a file, and discard the results from the foo command. I tried the following, but it didn't do what I want:
$ time ./foo > /dev/null >> output_from_time_command.txt
Following script name is 123.sh and I need to put this in the background if I do 123.sh -bg this will not bring me back to the prompt but echoes what ever I put (using echo hello >> /tmp/123) in to the /temp/123 file. the only way that I have found doing this is to do "nohup 123.sh &" to put this in to the background. Is this okay or is there any better way of doing this?
#!/bin/bash # file name is 123.sh tail -f /temp/123 | while read line
I am trying to use the time command to measure the execution time of a small program. The problem is that the command has three outputs. They look like this:
Code: $ time ./a.out real0m51.935s user0m51.060s sys0m0.040s
Should the execution time be the sum of the user and the sys time? sys time is really small.
I want to know how to set time format (12 hrs or 24 hrs) using command.I tried thisode:date +%T -s "2011-02-23 14:00"But it only displays 24 hrs format on TERMINAL but it does not SET 24 hrs format on the system
I would appreciate help with how to extract the date and time from at command jobs. From what I can tell, the date and time is embedded in the file name (/var/spool/atjobs).I'd be using this information in a (bash) shell script.
I use the time command to measure the wall-clock time of a GPU implementation of an algorithm. When I time the CPU execution of the algorithm time returns a negligible sys time. However, when I time the GPU execution time returns a sys time that is around 20-30% of the total time. If that time was comparable with the negligible sys time of the CPU I would achieve a speedup of a few times higher.
I suspect that the increased sys time is because of the GPU usage, which, I assume, takes some time for the OS because of the drivers etc. I am not sure though, and it is important to figure this out because it will improve my results a lot if I can ignore the sys time and use just the user time for speedup calculations. Also, is there a way to see, in detail, what is the sys running and takes so much time. I am thinking that I might be able to see if it is the driver indeed that causes this delay.
I was reading that if I want to do a one time scheduled command, I should use at, which I've never done, as opposed to cron, which i'm kinda familiar with. But what I want to do is reboot my server at 3am tomorrow and force it to check the file systems with a shutdown -rF. For this do I even need to use "at" or could I just say shutdown -rF 3:00.Will that also know that I mean 3am tomorrow and not say in 3 minutes from now or 3pm?
I am troubleshooting file copy time issues between 3 servers. I need to copy the same file from server A to both server B and server C and compare the elapsed times for the copy. Is there an easy way to do this?
So, I usually write/find a test case generator for any code that I write. This type of code generally leads to some file output. To be thorough, I try and generate many different files to test my code on.
Say the command is like this:
Is there a way to automate this for many different values of the parameters and generate many different files?
I tried:
I wasn't able to use the $i in the filename, and without it the command gave me no errors, but did nothing else either. I know the Unix command line is very powerful, and I have a feeling that this should be possible, but I just don't know how to do it.
while doing FTP through command prompt, connections times out after some time remaining idle, but doing ftp through browser, connection doesn't time out after some time remaining idle... why is it so?
I'm trying to find a proper command to move a certain set of files according to date/time range. I am thinking that the command should be something like:
I'd like to measure network latency for SNMP GET request. There is a free command line tool time which can be used to find timing statistics for various commands. For example it can be used with snmpget in the following way:$ time snmpget -v 2c -c public 192.168.1.3 .1.3.6.1.2.1.2.2.1.10.2IF-MIB::ifInOctets.2 = Counter32: 112857973real 0m0.162suser 0m0.069ssys 0m0.005sAccording to the manual, statistics conists of:
the elapsed real time between invocation and termination, the user CPU time (the sum of the