Fedora Installation :: Unable To Extact .rar Files - Error - There Is No Command Installed For RAR Archive Files
Jul 20, 2010
i am downloaded some e-books in the format .rar. when i am extracting them i am getting error as There is no command installed for RAR archive files. Do you want to search for a command to open this file?
I came across this feature of the "less" command today. It appears not to be well-documented so I'll share it here. I found that "less" can automatically uncompress compressed files - no need to run the first command, the second does the same job:
gunzip -c textfile.gz | less less textfile.gz
Even better, "less" can read and list the contents of some archive files as well, including Tar files compressed using Gzip and Bzip2 as well as Zip files. So now when I want to list the contents of an archive,for example to determine whether the contents will be extracted to an absolute path or relative to the working directory, the following work
less archive.tar.gz less archive.tgz less archive.tar.bz2 less archive.zip
Much easier than remembering the various options that the Tar, Gzip, Bzip2 and Zip use. The best one for me is with RPM archives:
The above outputs the Information, Change Log and verbose File List of the RPM (paging the output, of course!) The following commands (using short and long options) give the equivalent output using the "rpm" command:
rpm -qilvp --changelog some.rpm | less rpm --query --info --list --verbose --package --changelog some.rpm | less
"less" does not recognise the initramfs file produced by "dracut" (which is a compressed "cpio" archive), but I suppose you can't have everything ...
surprisingly come across this file that the ubuntu archive manager is simply unable to unpack.these are some articles that i am trying to download.strangely i have never heard of the format and the page asks me to download "stuffit expander" to allow me to view the pdf files. however when i download the file saving it on the desktop, and try to open it, archive manager is simply unable to do so!this is the title of the downloaded file -
i have a problem i have two ubuntu programs on my hard drive. i have old 9.4 and i installed new 9.10 in a partition. i was unable to get old 9.4 to work correctly and i installed new in fear of looseing pics and files. what do i do.
If I understand everything right I can use simple cd to create an install of my system then config everything and create an archive of /etc,/var,/usr and that give me a backup of my changed files. but what I am not able to do is create a script to replace the files with the archived version of the config files automatically. I need it to be ran as a post intall script or to configure the conf files as part of the install. what I configure on every reinstall. apache,proftpd,mysql,php,and wordpress. Since I am still learning I reinstall a lot and standard backups have not been working for me. But I can put the whole install including backup on one cd. Is there a way replace only the change files from the archive. or getting a list of just the modafied files on my system.
I have a .tar archive of music that I would like to extract specific files / folders from. I've been looking at man tar but I don't seem to be able to find any information related to my quest, although someone will likely point out that it's right there in front of my nose
Anyhow, the file is called music_archive.tar and I would like to extract the folder /music/P/Pink Floyd/The Wall/. Anyone know how to do this
It is saying to make sure there is enough room in the tmp folder, so I cleaned out the entire tmp folder but the same error is still happening. Is there a way to make the tmp folder bigger? Or perhaps extract the temporary files to a different folder? Or maybe the size if the tmp folder is not even the problem?
opensuse v11.2, linux 188.8.131.52-0.1-desktop x86_64 ZIP v2.32
I wish to exclude some files from a zip archive. On other OSes to exclude an entire directory I would use the "-x" option like so:
Code: zip -r archive-name * -x dir1/* Simple. And just add "-x"'s as needed (or use an exclusion file).
Not how it works here, it would seem. AFAICT all "-x" options are ignored. (The entries in an exclusion file also.) For instance, "-x diy/mplayer/*" should ignore everything in the <diy/mplayer> directory. It does not. I have tried fully qualified paths as well; no joy.
I download the live 64 and burned it to a disk after it burns to the disk Brasero checks and says "error some files could be corrupted." I burn it slowest burn same thing happen. I downloaded thinking I had a corrupted download. Same thing happen. I never had a problem burning iso until now. What am I doing wrong?
I am using the diff command with the -r option, to compare a large number of files and files in subdirectories. My main interest is to find out which files have been changed, and not what the actual changes are, and since a lot of files has been changed, it would be a lot easier to view the file names only. Is there and option for diff that might do this, or does there exist a similar tool/command that could do the job?
The problem I have is that I need to replace a more complex string, like this: Old string: /mnt/stor6-wc2-dfw1/627896/982574/ New string: /mnt/stor8-wc2-dfw1/369587/302589/ There I don't know how to do it... since the / is what separates the old from the new strings, and the strings that I want to replace have / in it. Also, I would like to know how to specify under what folder replace the files, for example, I want that it search/replaces all files under /var/www/mysite/htdocs folder.
Is there a method at the command line to copy files from one location to another and retain the source files group and user?I'm migrating some MySQL files from one machine to another.I want to back-up the original files in the directory presently. They have owner:group of mysql, some have owner:group root:mysql and so on. To copy them under cli or Nautilus everything changes to root for I execute sudo cp or gksudo nautilus and copy via gui.
Since it is MySQL data I could simply do a dump of the database and restore it on the other machine. But there's about 20 db's and I want to do this via a copy for it will be faster - at least that is what I think.
I have a RAR archive, split into 40 sections, containing a folder with a total of 2GB of data. When I Open the Archive using "File Roller 184.108.40.206" and drag the archived folder to destination folder, it says that its uncompressing, but after its uncompressed, there is nothing there, in the destination folder. Where was the 2GB extracted stuff gone?
I recently upgraded from FC12 to FC14 due to errors using yum and rpm. I keep running into mirror sites with RPM files that don't install. These site have what appear to be valid RPM files, but if I use wget to retrieve the file and run "file" on it, it turns out to be an HTML file. I understand the fact that if a file is corrupted, it won't install, but I can open these "RPM" files using Mozilla after changing the extension from RPM to HTML. When I open the new HTML file, I'm prompted by the browser to open or save the RPM file. If I save the RPM file pointed to by the HTML file, I can download and install it using rpm -ivvh unless a dependency check fails. how does anyone install using yum if the package is an HTML file with an RPM extension. Here is a link as an example:
If I use wget on this file and then run "file" on it, here is the output:
file openoffice.org-calc-core-3.3.0-20.2.fc14.i686.rpm openoffice.org-calc-core-3.3.0-20.2.fc14.i686.rpm: HTML document text
If I change the extension of this file to HTML and open it with my browser, I am prompted to download the RPM file which turns out to be a true RPM v3.0 file. Is anyone else having this problem or can anyone point me in the right direction to download the correct RPM file. I am running behind a proxy server and have installed cntlm, configured wgetrc, yum.conf and .curlrc to use the proxy server. I am able to download and install most of the simpler applications where the dependencies are either installed or non required.
Attempting to use PLACES>Home (or Documents,Bin,Desktop,etc.)brings up the error message "There is no Windows program configured to open this type of file." If I open Open Office I can find all files in home, documents, etc. and open them.The computer is otherwise functioning normally but using the usual methods all the files that should show look empty. How did this happen? I was attempting to use WINE to load Avery Wizard (a Win utility) through AWsetup.exe. To do this I changed permissions for that file but I have subsequently removed the file and its subfiles using the terminal.
I'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet: cd /media/"My Book"/"Linux Backups" NOW=$(date +"%b-%d-%y") LOGFILE=Backup_Root_FileSystem-$NOW.log sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
When trying to create a new compressed/archive file in Gnome Commander (GM) the file is created but the selected files are not added. I can open the new (empty) archive file and then add files to be compressed. I have tried using several different formats (zip, tar.bz and others) with the same results. The "file roller" is shown as a plugin but has no configuration other than the compressed file type.
AKA "zipping on the fly .. the slow-as-molasses way." The list includes full pathnames to each file, and they're all in subfolders of the same parent folder (which, unfortunately, is not the root folder of the drive or system on which the files reside). A cleaned-up and radio-ready portion of the list looks like
What I'd like to be able to do is zip all the files in the list into a single archive, to avoid the step of having to copy them to the same location (presumably another folder on the HD) and then zip that folder. I'm more inclined to make provisions about extracting to a single folder at some other time. Is this possible in BASH, or would I have to consider a faster, more robust scripting language such as python or perl?