General :: Extracting Latest Tar File From The List?
Jan 25, 2010
I have a bunch of tar files in a folder with the name in 'file_name.MMDDYYYY.HH.mm.SS.tar' format. I need to write a shell script that picks latest and greatest of the above tar files and extracts it to a particular folder
If I go to the website and download the ISO file is it the exact same one that was released in April or will it contain the latest patches (e.g. security fixes, etc) that have been release since then? If it doesn't I suppose I'll have to download the entire ISO file, install the OS, and then have another couple hundred MB to download which is not what I'd prefer.
I have a remote Debian 5 machine on which I'm trying to extract a tar-file (Django 1.2.1, but that's unimportant). The first 1300 or so files extracts just fine (of about 2800 in total), after which errors liketar: Django-1.2.1/tests/urls.py: Cannot open: No such file or directoryshows up for the rest of the files, finally ending up with atar: Error exit delayed from previous errorsI use the commandtar xzvf Django-1.2.1.tar.gzI've also tried with sudo (but I'm extracting in my home directory, which I should obviusly have permissions to use without). I've tried gunzipping first (runs flawlessly) and then running tar xvf, but to no avail. I've fetched the file using wget:
I am really trying to learn the heart of linux, and not rely on gui's and package managers for everything. I am trying to download firefox 3.6. I downloaded the arc file, and extracted it to my home folder, but I don't know what to do next. I am also clueless as to how to install tar or tar.gz files, and tell them where to go, etc. Firefox is the priority question here. I am sure I can get it from the package manager, but I want to learn to do it with the arc files, and tar files, etc, and not rely on the gui for everything like in Windows. I am a Linux newbie, converting to Linux for hobby and educational purposes, but I have no direction, or no sense of where to go, or how to learn. I am running Kubuntu 10.04 LTS.
There are some log files that I wish to get some information from (Apache Access Log) but it is huge! All I need as of right now is any information from date and time A to date and time B. What commands can I use to extract this information from the access_log and put it into another file with just that information? I created a file called "access_info" by doing Code: touch access_info but I was not sure where to go from there.
I am trying to create a self extracting file for Windows from Ubuntu 10.04 using 7zip.
I tried these commands:
7zr -sfx7z.sfx a output.exe *.txt 7zr -sfx/usr/local/etc/7z.sfx a output.exe *.txt
Both show this error:
Error: can't find specified sfx module
System error: E_FAIL
I have all the .sfx modules on /home/username/.sfx/ because I use them previously with "rar" and it worked fine, but it's not working for 7zip. I also copied 7z.sfx to /usr/local/etc/ to test the second command like shown here but I get the same error.
My OS is Fedora 14 - 64bit. I want to extract and install this file "LEXMARK-INKJET-LEGACY-wJRE-1.0-1.x86_64.rpm.sh.tar.gz". At the moment this is in the download box in the top left corner of my monitor.
I have tried several ways, but all I get is Command Line Output. Gzip:stdin: unexpected end of file./bin/gtar: unexpcted EOF archive. /bin/gtar: error is not recoverable: exiting now.
I have this file that is in ZIP. I can open it and browse the files inside via archive manager. But extracting it gives me an error. When using GUI it says that 'ERROR while extracting files' and in command line when I command: unzip -q myfile.zip
It display this errors
file #2109: bad zipfile offset (local header sig): -2122072556 file #2110: bad zipfile offset (local header sig): -2121511988 file #2111: bad zipfile offset (local header sig): -2121511886 file #2112: bad zipfile offset (local header sig): -2121375823 file #2113: bad zipfile offset (local header sig): -2121310073 file #2114: bad zipfile offset (local header sig): -2121282474 file #2115: bad zipfile offset (local header sig): -2121107851 file #2116: bad zipfile offset (local header sig): -2121015601 file #2117: bad zipfile offset (local header sig): -2120957317 file #2118: bad zipfile offset (local header sig): -2120906286 file #2119: bad zipfile offset (local header sig): -2120906193
I'm trying to extract a file from a zipped archive, and receive an error message saying "caution: filename not matched:".
I suspect the route cause of this is that the original filename contained accented characters (which now show as question marks in the zipped filename). I've tried renaming the file in the archive, but get the same error.
I am trying to extract data from the xml file using c coding. here is the sample code. this code gets compiled with errorfree. but the call back functions start and end are not invoked. coding:
i want to get the latest file name from a directory, how can i do with perl script. In other words, how to sort the contents of a folder on the basis of time and capture the latest file using perl script,
ish but I did manage to right-click the Firefox tar.bz2 file and choose to 'extract here'...now I've got the new extracted folder on the desktop but I'm lost as to how to run the installation of the 3.7 alpha 4 release.
9.10 Wubi with XP until 10.04 Google Chrome Widmer Drifter Pale Ale Tommyknocker Pick Axe Pale Ale 1992 Davies Washburn N4 w/ Suhr SSV/DSH+ pups Jazz and Prog Rock, Dream Theatre & Porcupine Tree
I want to write a script. I have the basic form working... but I would like to add some advanced features. I have a list of .zip files that have a version number in the filename (i.e. sudoku-134.zip, sudoku-176.zip, sudoku-203.zip, etc). I would like to have a script that will find the latest version of each game, and unzip it. I want the next to latest version left untouched. But anything older, I would like the unzipped folder deleted (But leave the .zip file)
Here's what I have so far: Code: #!/bin/bash cd /var/www/vhosts/server.com/httpsdocs/games/$1 for file in `ls *.zip`; do if [ -d ${file%.*} ]; then echo else mkdir ${file%.*} cd ${file%.*} unzip ../$file > /dev/null cd .. echo 'date' $file >> /var/www/vhosts/games4roku.com/private/$1.log fi done I can call the script with "game-unzip public" to unzip all the public games.
Here's my files: Code: sudoku-134.zip sudoku-134/ sudoku-176.zip sudoku-176/ When I add the latest version, sudoku-203.zip, I want it to unzip it to ./sudoku-203/ and rm -rf sudoku-134
About a couple of days ago I'm sure I saw that Open Office was offering a DVD file download of their latest version. It was including versions for Win. Mac and Linux all on the one DVD. Now, after buying a DVD burner for my 'puter, I have been unable to find that information again.
I'm trying to extract the audio from an mp4 video file and convert it to mp3 using ffmpeg. The problem is that the resulting audio file is only 3m2s long, whereas the video file is 8m15s long. Anyone know why the audio file is being truncated? Am I doing something wrong? Apart from being truncated the mp3 file plays perfectly. Here's what ffmpeg has to say about the video file:
########## some text text also includes empty lines ########## some more text ##########
Basically all sections are separated by 10 hashes and I need to somehow only print all lines in the last section (the "some more text" part in the example above"). I tried all kind of things with sed and awk but I didn't find any way to identify the last "section".
I use Lenny, and was trying to mount a .iso image, supposedly a cd imagem.
[code]....
This is what I get from dmesg | tail:
debian:/home/zac/cscd# dmesg | tail [ 1811.505199] floppy0: disk absent or changed during operation [ 1811.505207] end_request: I/O error, dev fd0, sector 0
[code]....
I did a little research on the web and it seems that this file is not really a cd image, but simply data in a .img file. What do you think of that?
debian:/home/zac/cscd# file cscd3.iso cscd3.iso: data
Some people recommend to extract the data via the dd command, but it didn't seem very safe for me to do that!
[URL]
is it possible to extract the data into a directory (instead of a device) using dd? This file is supposed to be a software. I wanted to run it on wine by keeping it mounted on a mount point in my file system. Does it make any sense to try to do this if the file simply isn't a cd image?
I am trying to extract just a few files/folders from a Time Machine backup, but can't seem to find them. The drive is automatically mounted in Ubuntu, and am able to access after enabling view hidden files, the HFS+ Private Data Directory. But that is a jumbled array of thousands of numbered folders, with each taking a fair amount of time to open on this aging Dell running 10.10.
I've tried running the standard Places - Search for Documents, with 'Show hidden and backup files' enabled, but that won't pull up any of the search times I'm going for (and seemingly won't find anything at all on the drive). So, is there any way to decipher the directory tree so as to be able to access this data from Ubuntu? Or perhaps a file embedded somewhere in there that lists out the original structure, so that I can use it as an index to see what number correlates to what originally named folder?
I have generated a list of directories that I would like to use ls and grep on, but it is not working. I am using the commandCode:cat directories.dat | xargs lsand I get a whole lot of these errors:Code:ls: cannot access ./foo/bar/baz/grault/*: No such file or directorybut when I try the directories manually one at a time I find that they all exist and all have files in them. Same thing if I try to grep anything. What is going wrong?
I've published a samba share on a Ubuntu 8.10 server, and mounted the share on a windows box. For some strange reason the files are not sorted by file name (nor by any other sorting type) if I list them from the Windows command line (dir). I could always use the "dir /on" parameter, but I hope there's a way to make samba sort the files for me
I am trying to write a script to pick the directory name from a list of file. Here is a detailed picture.Have a file name LIST which contains the follwing for example/apps/oracle/product/test1/apps/oracle/product/test2/apps/oracle/product/test3I need a script that reads these line from LIST and creates foldersin /apps/oracle/product/test1/backup/date/test1 after reading the first line /backup/date/test2 after readin the second line/backup/date/test3 and so on.
I am using endian firewall 2.4 and squid 2.6, everything is working fine for me. In squid report i am getting download list by user or IP based. Now my question is, is this possible to get user based uploaded file using HTTP POST. Is there any configuration have to be done is squid.conf, if answer is yes then please give me the what is configuration i have to give. Example: if suppose user uploading a image file into facebook.com, i want to track the file name and which IP address and then where its uploaded.
I have a system that will not boot as /usr has been destroyed and I would like to get a list of installed packages before re-installing. I know that it's possible to get this using dpkg or apt, but I cannot run those.
Where in the filesystem is this information stored and what's the best way to get a list of installed apps from the files?
With file command we can get the file type. File returns a unique answer for each file type. I want to get the total list that file can return for linux & solaris. For example, I have a partial list of file's return output. Unfortunately I can't see the total list from manual page.
ASCII C program text ASCII C++ program text ASCII C program text, with CRLF ASCII C program text, with very ASCII English text ASCII English text, with CRLF line
I have a text file with a long list of RPM's. I need to check if each RPM is installed. I'm sure I can cat out this file and run "rpm -qa" against it, but I'm having trouble with the syntax right now...
It's always a good to backup a configuration file like sources.list before you edit it. To do so, issue the following command: sudo cp /etc/apt/sources.list /etc/apt/sources.list.backup Where does it backup to and how do I access it? I want to put the backup on removable disk or upload it