I'd like to write a script that invokes a gnome-terminal session which slowly reads out text like the phosphor screensaver (could be anything, a log file, ascii art, song lyrics, whatev) and then closes. I can invoke a terminal using [gnome-terminal -e 'cat /var/log/dmesg'] but the output flies pass by too quickly.any way to slow it down? I know it seems like an odd request but if anyone has a suggestion I'd love to hear it.
I am trying to find sed command combination to print out the "start command" line, the id line and all lines between "details" and "stop command" only if "error" exists. Here's the original output (test.txt):
im trying to output a list of running processes via a shell script. At the moment i got this which outputs the processes to a text file called out.
echo $(ps aux) >>out
The problem is though, the processes are all just one big block of text which makes it hard to read. Does anyone know how to sort the output to a text file so that it prints to the text file at 1 process per line? I know its probably simple but im very new to linux.
hello I tried to find a good subject but it was the best of mine, anyway I'll explain it here. some time I do some thing like installing a new application in Linux terminal of my office PC but it take a long time and I have to go home during its installation or configuration process that it is not good to cancel it.My current solution is abandoning the process until next day. I wanted to know is there any way to redirect an input and out put of a terminal to another one, if it works I can continue my abandoned process by ssh to my Linux office PC and redirect that terminal to my new remote sshed terminal from my home.
if you do the command conky in terminal, it starts conky ofcourse, but it also shows output to that terminal so you can't do any other commands to that terminal, Is their an option like you can do with the '&' sign in other cases? If you do the '&' sign with conky it still gives output, also the conky -d command gives output...
I was wondering if it is possible to append some text to the output of ls. Like say, if i wanted to create symbolic links for all the files under a folder in my hard disk to a folder on my desktop, I could say (Pretty sure this won't work, but I am looking forward to something like this) echo ln -s | ls . This should append ln -s to all the files of ls.
I'm having a slight dilemma on reading data from a text file and outputting it into a table then displaying it. Basically I'm writing a shell script that takes information from text files then outputs the data into a table with 4 headings. The extracting of the data is fine, but creating a table i'm having problems with. My code extracts the data outputs the string to another file which works fineThe text file looks like this
mr smith 1 purchase oct 2007 mrs smith 2 purchase nov 2006 i want it to look like this
Whenever I start a text editor, such as kate, kwrite, gedit, it takes at least 10 seconds to start up.This on a PC with 4 cores and 3GB RAM.I know that these text editors should start up much faster, like, I click the launch icon and it should already by there.But instead it's 10 seconds waiting and then the editor is there. Why, what's it doing during those 10 seconds, and most important, how can I fix this?
has no MTA (Postfix, Sendmail, Exim) installed so it can't email me the results of any cron job I schedule. I would then like to have the results from Cron be dumped into a small text file so I can read it later to view any issues.Right now my job is as follows:Code:01 18 * * * /usr/bin/shellscript.shAbove you see my script I want to run every day @ 6:01 pm. My question is what would the cron line look like if I wanted the results dumped into a random text file somewhere on my system?
Is there a way to color particular words printed on console based on user preference? For example I need to color text 'error' when a particular program is compiled.
it compares two files using md5... if they are same , a corresponding character is output to a text file .. but the problem is it gets appended by default.. is there any way to output in a normal way because the text is a message and it should be of proper format here is my script
Code:
#!/bin/bash g=`tail -1 new.txt|head -n 1` array=( a b c d e f g h i j k l m n o p q r s t u v w x y z ) for((i=1 ; i <$g+1 ; i++))
[code]....
the message is supposed to be hello , i need to get rid of the endlines somehow..
just upgraded to Lucid. When booting, I get a blinking cursor on a black screen for 30 sec. Then I get some flashes of text about UDEV. Finally, I get a desktop after about 30 more sec.
Any way to speed this up? I've already tried disabling my nonexistent floppy drive in the bios, but it didn't help.
I would like to capture all output spewed to a terminal session including processes that are terminated that were invoked from a script running in a terminal window. this is beyond capturing just stderr and stdout . for example
{ ./script } 2> stderr.cap 1>stdout.cap
if script is terminated (including because of memory violations) I get spewed output to the terminal I would like to capture that spewing to a file automatically or to a bit bucket /dev/null Is there another filehandle which can be redirected to do this? If so how or is there another way???
In synaptic package manager when I expand "Details" tab (at the time when some software is installed or removed), I can see a terminal. I want to capture the output of terminal into a file. I tried to highlight and right click. But there is no context menu(copy, cut)
I sometimes stick my neck out and provide somewhat detailed, and often risky, "Mr-fix-it" remedies for boot problems. Now, I know it's possible to amend each command with "whatever_command > whatever.txt" in which case it'll place the command output in a file in /home.
But if you're directing someone to run a lot of commands as I did here is it possible to save the output of all commands to a .txt file without amending each command?
Or is it already saved somewhere that I'm not yet aware of? I wouldn't be surprised if the latter were true, I just haven't yet found it
I'm having a slight dilemma on reading data from a text file and outputting it into a table then displaying it. Basically I'm writing a shell script that takes information from text files then outputs the data into a table with 4 headings.he extracting of the data is fine, but creating a table i'm having problems with. I think it is possible to do it using the awk function, but so far i'm having a lot of difficulties.
Am having issues getting the output from a script to be logged in a file. I need the script to output both the stderr and stdout to the same text file.
For some strange reason my gnome-terminal has become really slow and unresponsive.If I type anything on the terminal I have to wait a few seconds before anything appears on the terminal.Memory or cpu is not the problem. Currently I am using only about 20% of my CPU:s and 15% of my RAM..Booting didn't help.Lucky this doesnt happen in openoffice, firefox or gedit. (which I need most)and I am running 9.10, which has last been updated about a week ago.
I've been using openoffice.org a lot recently and I was hoping that I'd be able to cat the contents of the file solely through the terminal.What's the best way to do this? My thought is to use a command in a terminal to convert to plain text and cat the file.
I'm attempting to write a bash script to retrieve information from a websitemy problem is I can't seem to figure out how to simply copy the text from the website (I have tried multiple text-only browsers but they all seem to require some input to be able to do this where I'm looking for it to be done silentlyThe purpose of this is I have created a google doc that is a bash script itself and I want the one I'm writing to download this bash script and execute it (the original will be hard coded on a cd and I'm doing it this way for easy updating) only thing is when I use wget it downloads the html file with the script embedded in it and not the script itself
I want to write expdp output in a text file using a shell script
If i write like below:
It will write whatever is there in log file to text file
But, sometimes export fails with out start taking export (without generating log file) because of job already exists error. such times, we dont know about that error until we check manually... so i wrote like below:
But still it is not writing anything in to text file using above stmt...