I am in the process of moving from a windows to a linux box. I zipped a folder of images on the windows box and ftpd it to the linux box. When I try to unzip the file I get:
i have downloaded the software bluez-4.81.tar.gz from internet and saved iy on desk top.in command line i typed like following.
george@george-desktop ~ $ tar xvzf file:///home/george/Desktop/bluez-4.81.tar.gz but i have got result like this... rsh: Could not resolve hostname file: Name or service not known tar: file:///home/george/Desktop/bluez-4.81.tar.gz: Cannot open: Input/output error
[Code].....
there is a option for unzipping graphically.so i right cliccked using mouse and extracted.then a folder named "bluez-4.81" appeared on desk top. Now what are the steps to install this bluetooth software?
I want to extract a huge .tar.gz file but when I do extract it stalls the server. The server is write heavy and extracting seems to choke the disk. Is there a nice way to extract without stopping the world? I've tried the 'nice' and 'cpulimit' command but they don't seem to do the trick.
I'm looking for a program that will digitally display the time in three different cities - all showing at once. I don't care if it is a panel applet or stand-alone. I'm using Suse 11.3 and Gnome.
I'm in Gnome but if I logout I can get to a menu where I can choose KDE as well as other window managers. The problem is I have a program that's running inside Gnome and I don't want to stop it. Is there some way I can get into KDE without having to stop this program?
So, I usually write/find a test case generator for any code that I write. This type of code generally leads to some file output. To be thorough, I try and generate many different files to test my code on.
Say the command is like this:
Is there a way to automate this for many different values of the parameters and generate many different files?
I tried:
I wasn't able to use the $i in the filename, and without it the command gave me no errors, but did nothing else either. I know the Unix command line is very powerful, and I have a feeling that this should be possible, but I just don't know how to do it.
I am fairly new to Linux and was needing some help on a comparing more than 2 files. I am try to come up with something that would compare at least 10+ different files to a master file and give me an output of what is missing.
Example would be: a.txt, b.txt, c.txt, d.txt compare each of them to the master.txt file, than output the missing text for each file into new file.
I came across comm and diff commands, am I looking in the right place or is there a much easier way of doing this?
I have a directory with hundreds of html files. For all the files I have to: - delete all the row from the beginning of the file to the sentence "<img src="immagini/_navDxBottom.gif" />".
- delete all the rows from the sentence "<br clear="right" />" to the end of the file.
I often use the rpl command to make changes to multiple html files at once. For example:
rpl -R '<br />' '<br /><br />' mydirectory However, I haven't been able to figure out how to change multiple lines. For example, let's say I want to change all occurrences of :
How do you join multiple MP3 files into one? "cat" and "mp3wrap" are no good as they produce non standard MP3 files. I know I can use audacity, but when you have 1000's of MP3 files to join into one, it takes too long.
I would like to find all the files that contains the strings I'm searching.
For example (it's just an example), I would like to search all the files in "/etc" that contains "eth0" and "us", whatever where are located those 2 strings, the important is that the 2 strings are in the files listed.
It would be something like a "grep -lr 'eth0' *" and "grep -lr 'us' *" but in one time/command, so that I don't have to make a comparison of the 2 list of files resulting from the 2 "grep" commands given higher.
I have a file with 5 columns. Column 4 contains numbers.Is it possible to split the file into multiple files using a condition for the contents of column 4 i.e if column 4 contains a value between 0-10 then print the lines to a new file called less_than_10.txt
I have 60+ directory's each containing multiple .doc files. I need to move them to a single directory and keep their file name intact. I don't think cp will do that with out listing all the file names. I was thinking of something like: cp -r /dir/*.doc /newdir . Or should I use a combo like find -type *.doc|cp /newdir?