I put a text file on my desktop and added a couple lines of text with gedit. File type shows text/plain. Double-click opens the file in gedit which is what I want. I'm using the file to temporarily hold some snips of code that I copy from file to file, but when I copy some html into the file and save it, now file properties show it's text/html and a double-click opens the file in firefox, which isn't what I want. Is there some way to keep the file type from changing itself?
I need to be able to convert HTML email messages saved as text files (.eml or .msg) to PDF documents, one PDF per email, retaining formatting and images.
Are there any Linux tools that will allow me to do this from the command line (so it can be scripted)?
i'm trying to convert a html file into a text file when i simply run "html2text <filename>" the output displayed is the way we want but when i redirect the same using "-o" or ">>" the file is having extra characters in it. i even tried -ascii,but no much use.
I am looking for some source package which will convert html fileto text file without using perl.I mainly need to do this on an ARM platform, so if I get sources I can cross compile it. I need the source package in C
When I print receipts from certain company's emails, they span three pages or more bloated with images, links and overly large text. I can view the message in plain text by setting the preferences to only display text instead of html, but it doesn't affect what is printed. I still get the full HTML page printed. Is there a way to force evolution to print in plain text instead of HTML?
Basically, I have a selection of .rdp shortcuts to various machines, but they are considered the file type "plain text document (text/plain)" so if I change it to open with to tsclient, all of that file type opens with tsclient. And if I change it to gedit, all the rdp files open with gedit. Is there a way to create a custom file type dependent on the extension rdp, or is there a way to set tsclient to open by extension rather than filetype?
I'm trying to run a CGI file with Apache2, but when I navigate to it, I just get the file in it's plain text format and not actually parsing the file. What do I need to configure?
I've tried this Code: <Directory /var/www/> AddHandler cgi-script *.cgi Options +ExecCGI </Directory> And I've tried this Code: <Directory /var/www/> Options Indexes FollowSymLinks Includes ExecCGI AllowOverride All Order allow,deny [Code]...
i have a big file of random numbers i generated at some point in time, after working with it with different things(how fun that was)... i want to remove duplicate lines and i'm not sure i'm doing this right
I would like to know if it would be easy for me to get a program which, given a plain text file as input, discards the line separators and writes the rest as its output. And let's put in the case of a file that has been mangled to the point of having CR,LF (carriage return, line feed) in some places, only CR in others, and only LF in still other places. That is, the three possible combinations used in systems as a newline char (the 1st is, or was used by m$- dos, the 3rd one by Unix and I know systems where CR is the line terminator).
After all, all the program has to do is, every time it finds a char belonging to the set {CR, LF}, cast it away.
I just downloaded a Plain Text File that contains (what appears to me as) code that will 'unlock' functions of my MS keyboard, now what do I do with it? Here is a preview of some of the contents of the file from the beginning of the text.
[Code]....
.....and so on with different items. My goal in doing this was to enable the zoom button on the keyboard.
Some more HTML code... I would like to cut the above text so i get this: Sometext on multiple lines like this.Sometext on multiple lines like this.Sometext on multiple lines like this. Sometext on multiple lines like this.Sometext on multiple lines like this.
There are other HTML files with similar cuts I need to do, but once I have the method for doing one, I am sure I can do the others.
I think the two logical strings to cut between would be:
I am not sure if these strings are always the start and end of the line respectively, is this makes a lot of difference! Then the HTML tags would need to be stripped to get the text on its own.
I know the commands for removing tags, but searching for a string like class="IOSSectionTitle", and cutting everything before it etc is something I am finding challenging.
Just thought I would add that the HTML does not nec. appear on logical new lines throughout the file and there may be unexpected new lines, but as far as i know the class="IOSSectionTitle" and <img always appears as a string without any new lines between those characters.
i working with a simulator tool that i need to pass to it a file in .BIN format, basically i need to convert from a tex plain file to BIN file How can i do that? there is some command(s) that allow me do
1) capturing an RDF formatted RSS feed as a file on my computer
2) converting the result to HTML using local command line tools
I've sorted 1) with wget? I've discovered xsltproc but I'm going round in circles. The master plan is to import my pinboard bookmarks into a static web site produced on my linux box using a handful of clever bash scripts.
I need to extract som text from a text file. The text is a test log with system info at the top and results further down. What I need is to add different tags with formatting before and after each line. I have prepared a template with html formatting, but the number of lines in the test log may be different from case to case, so I need to be able to add formatting tags by need. Can this be done using bash script, sed, awk, head, tail... ?
I use the below perl program csv2xls.pl can convert a csv to xls file , it works fine , but I found that it only work for all English characters , I tried to use it to convert a csv file ( with Japanese characters ) , it does work , I also tried the perl "unicode_utf16_japan.pl" , it also did not work , can advise what can i do ?
I have 2 GB of html files and i need to convert all of them at once to text files. Is there any way that I can do it. Converting each of them manually is really a pain and I want to convert all of them with a single command to text files.
I have a few problem. I have txt file like this:Quote:00 21 55 84 9a ff 00 1f 9e 1a 5b 00 08 00 45 00 00 4b 00 00 40 00 3f 11 9a 0e a1 8b fa 02 04 02Then, based on my txt file, I would like to generate text like this:Quote:00215584 2155849a 55849aff 849aff00 9aff001f ff001f9e 001f9e1a 1f9e1a5b 9e1a5b00 1a5b0008 5b000800 00080045 08004500 00450000 00004b00 004b0000 4b000040 00004000 0040003f 40003f11 003f119a 3f119a0e 119a0ea1 9a0ea18b 0ea18bfa a18bfa02ased in my reading, I found about ngram solution in perlbut I not really understand to edit from source code given. I m begineer user in programming language. I hope to get the solution. [URL]
I need to convert a very large latex project (made up of many .tex and style files) into .html (or something similarly non-.pdf). Can someone recommend a quality converter program? Preferably, one that is:
on creating a new perl script which replace IP address from the text file. eg. If in a file, we found any word like 11.222.333.44 then it has to be replaced to XX.XXX.333.44
I have an pdf file on my linux RHEL 4.7 machine. I can open that file but when i click on 'saveas' to save the file in 'Text' format there are no options i see there. I need to save the 'pdf' file to 'text' format. could anyone tell me how to save the pdf file to Text format. Iam using 'KDE'
I have a gedit text file 2.2MB. I want to convert it into two orhree smaller files/volumes, so I can upload them seperately to web pages. Does anyone know a quick and easy way to do this?
i am collecting usb usage details of all users and convert it into csv files so that i can export it into some database..the output desirable is in csv format for database with some batch or awk script.