I have come across a webpage where there are several (thousands of hyperlinks). Of those hyperlinks, many are named "CDS" and in each of these CDS hyperlink- there are certain lines which i need to chop off!! I kind of have this feeling that AWK would be the one to use.
I use Lenny, and was trying to mount a .iso image, supposedly a cd imagem.
This is what I get from dmesg | tail:
debian:/home/zac/cscd# dmesg | tail [ 1811.505199] floppy0: disk absent or changed during operation [ 1811.505207] end_request: I/O error, dev fd0, sector 0
I did a little research on the web and it seems that this file is not really a cd image, but simply data in a .img file. What do you think of that?
debian:/home/zac/cscd# file cscd3.iso cscd3.iso: data
Some people recommend to extract the data via the dd command, but it didn't seem very safe for me to do that!
is it possible to extract the data into a directory (instead of a device) using dd? This file is supposed to be a software. I wanted to run it on wine by keeping it mounted on a mount point in my file system. Does it make any sense to try to do this if the file simply isn't a cd image?
I am trying to develop a method of reading files generated by other programs. I am trying to find the most versatile approach. I have been trying bash, and have been making good progress with sed, however I was wondering if there was a "standard" approach to this sort of thing. The main features I would like to implement concern reading finding strings based on various forms of context and storing them to variables and/or arrays. Here are the most general tasks:
a) Read the first word(or floating point) that comes after a given string (solved in another thread)
b) Read the nth line after a given string
c) Read all text between two given strings
d) Save the output of task a), task b) or task c) (above) into an array if the "given string(s)" is/are not unique.
e)Read text between two non-unique strings i.e. text between the nth occurrence of string1 and the mth occurrence of string2
As far as I can tell, those five scripts should be able to parse just about any text pattern. I am by no means fluent in these languages. But I could use a starting point. My main concern is speed. I intend to use these scripts in a program that reads and writes hundreds of input and output files--each with a different value of some parameter(s).
The files will most likely be no more than a few dozen lines, but I can think of some applications that could generate a few hundred lines. I have the input file generator down pretty well. Parsing the output is quite a bit trickier. And, of course, the option for parallelization will be very desirable for many practical applications.
I have been saving an XML file wherein I shall be reading data, if and only if the XML is present on Hard Drive. If same is not present, the code shall generate first and next time it shall simply read and not jump into writing part again and again. The issue that I am facing is, I am deleting the file manually going to that location, and checking whether it is reading the data, upon investigation, I found that it is really reading the data although the file is deleted.
I suspected that there may be some cache issue from where it is picking values, hence it fired the command "sync; echo 3 > /proc/sys/vm/drop_caches" But it is still doing the same pattern.
streaming videos i sometime want to save them, after watching. generally they're saved as flash videos or mpeg-4 video, which is all fine.however periodically they're split-up and saved in smaller chunks with names like data_1, data_2, data_3 etc. these range from 14.0 - 44.0 MB. the file in question (currently i'm trying to save from the cache) was from divxden, or possibly divxstage.eu either way i think it used the totem plugin.so, my question is: does anyone know if these files can be stuck back together, or if this feature can be changed so streamed files are kept intact?
I was laughing about klackenfus's post with the ancient RH install, and then work has me dig up an old server that has been out of use for some time. It has some proprietary binaries installed that intentionally tries to hide files to prevent copying (and we are no longer paying for support or have install binaries), so a clean install is not preferable.
Basically it has been out of commission for so long, that the apt-get upgrade DL is larger than the /var partition (apt caches to /var/cache/apt/archives).
I can upgrade the bigger packages manually until I get under the threshold, but then I learn nothing new. So I'm curious if I can redirect the cache of apt to a specified folder either on the command line or via a config setting?
I installed squid cache on my ubuntu server 10.10 and it is work fine but i want to know how to make it cache all files like .exe .mp3 .avi ....etc. and the other thing i want to know is how to make my client take the files from the cache in the full speed. since am using mikrotik system to use pppoe for clients and i match it with my ubuntu squid
i was looking for a way to stop my menus taking a few seconds to load my icons when i first open them and found a few guides suggesting using the gtk-upate-icon-cache command, but with the any colour you like icon theme i'm using (stored in my home folder .icons directory) i kept getting a "gtk-update-icon-cache: The generated cache was invalid." fault i used the inbuilt facility in the acyl script to copy the icons to the usr/share/icons directory and tried the command again, this time using sudo gtk-update-icon-cache --force --ignore-theme-index /usr/share/icons/ACYL_Icon_Theme_0.8.1/ but i still get the same error. i tried with several of the custom icon themes i've installed and only 1 of the first 7 or 8 i tried successfully created the cache.
I've just downloaded a vidio from ..... and converted it to an .avi file. What I want, however, is just the soundtrack as an .mpg to add to my mp3 player. I'm sure Fedora has software to do this, but I don't know enough about audio/video editing to know where to look.
My OS is Fedora 14 - 64bit. I want to extract and install this file "LEXMARK-INKJET-LEGACY-wJRE-1.0-1.x86_64.rpm.sh.tar.gz". At the moment this is in the download box in the top left corner of my monitor.
I have tried several ways, but all I get is Command Line Output. Gzip:stdin: unexpected end of file./bin/gtar: unexpcted EOF archive. /bin/gtar: error is not recoverable: exiting now.
I never had any problems extracting any rar files until about a month or so ago. Now I can't seem to extract ANY video file that has been rared. It is becoming a pain to have to boot into windows just to perform a 15 second task. I dont have any problems with any other files except video files. I get a 'CRC Failed' message. None of my searches have yielded any results. Does anybody know what is causing this problem and how to correct it? Or yet, why it just started happing? I'm using karmic, and was using p7zip-rar until the problem began. The I added unrar and unrar-free but that didn't make any difference.
I created a tar of the /var/www/html directory from a live web server and I want to extract it onto another server that will eventually become a backup server to the live server.When I extract the tar file, I see the whole directory structure /var/www/html. How do I extract the file just leaving the contents of the html directory in the directory of the new server?
After having downloaded unrar, I find I can't actually use it to extract rar files. Whenever I use the command unrar e filename.rar, (obviously using the correct filename), I just get told there's no such file or directory.
I am using XFCE spin Fedora, so my default archiver is Xarchiver. I have downloaded a 2 part .RAR file, and when I used to use them in Windows, I would tell WinRAR to extract the first part, and then it would automatically start extracting part 2 if it was in the same directory. When I try to do the same in Xarchiver it tells me "An error occurred! Please check the 'Store archiver output' option to see it." I don't know what 'store archiver output' is or how to access it. What do I need to do to extract both components and successfully join them?
I have a large number of paper sheets to scan. Instead of an autofeed or flatbed scanner, I'd like to use my photo camera. After a while, it occurred to me that instead of making a photo for each page, I could install the camera on a tripod, set it to record video and then extract the images from the video.
That way, I'd just have to handle the papers: Pick one, place it in front of the camera, wait a moment, replace it. Is there a Linux software to get all images without motion out of a movie? Kind of a non-motion-detection.
I have this file that is in ZIP. I can open it and browse the files inside via archive manager. But extracting it gives me an error. When using GUI it says that 'ERROR while extracting files' and in command line when I command: unzip -q myfile.zip
It display this errors
file #2109: bad zipfile offset (local header sig): -2122072556 file #2110: bad zipfile offset (local header sig): -2121511988 file #2111: bad zipfile offset (local header sig): -2121511886 file #2112: bad zipfile offset (local header sig): -2121375823 file #2113: bad zipfile offset (local header sig): -2121310073 file #2114: bad zipfile offset (local header sig): -2121282474 file #2115: bad zipfile offset (local header sig): -2121107851 file #2116: bad zipfile offset (local header sig): -2121015601 file #2117: bad zipfile offset (local header sig): -2120957317 file #2118: bad zipfile offset (local header sig): -2120906286 file #2119: bad zipfile offset (local header sig): -2120906193
I have a bunch of split Z files and I want to extract them but I cannot find how. To be more specific I have a file named foo which is compressed into files foo.0Z, foo.1Z foo.2Z. I have tried uncompress but does not work I have tried gunzip,gzip etc but no result either. Even 7zip will not do the work.
I am trying to install MySQL, Apache, and PHP onto my Linux Ubuntu laptop from a cdrom that I got from a teach yourself MySQL, Apache, and PHP book I bought. First off there were two steps which i think i did right. The first was to mount the CD-ROM under /mnt in my file system which i did and there seems to be no problem except for the reply "mount: block device /dev/sr0 is write-protected, mounting read-only"
now first of all there are two things I need to explain, first of all I was superuser for all of this by using the command su at the beginning of my terminal session. Second of all the symbol "|" on my line of code above looks different from the symbol in the book. It should have a small space in the middle of it. However it looks like this on my keyboard so I am assuming it is simply a difference in text font. That is it, it would be really nice to figure out what went wrong because to be honest I am curious as to why the book is giving me code that gives multiple errors.
I'm trying to extract a file from a zipped archive, and receive an error message saying "caution: filename not matched:".
I suspect the route cause of this is that the original filename contained accented characters (which now show as question marks in the zipped filename). I've tried renaming the file in the archive, but get the same error.
I recently saved a large number of web pages from a website on my computer and all of them contains images. I needed help extracting all the images from the webpages(all of them) there are about 17000 saved webpages in the folder and all of them have images. I am not sure how to extract images from webpage in batch. google search dint turn up anything.. is there lika a tool or something for the job or should a script be written for it?
I have a PHP script written that is checking a string to see if it contains a link in it (i.e. a URL). I have the following if statement, that uses 3 possible regular expressions to determine if there is a link or not.
Code: // check if we found a link // links are denoted by strings that: // - contain http:// // - contain www.*.*
I'm not convinced yet that writing a shell script to do this is the best course of action. If someone is capable of doing this with a Perl or a Python script that's fine too. If you want to make it super high performance and write it in assembly