Software :: Uncompress Zipx Files?
Jun 11, 2010I have been given a 1.9GB zipx file (created with a recent version of winzip). I can't find a way (and yes I googled it) to open it on Linux.
View 2 RepliesI have been given a 1.9GB zipx file (created with a recent version of winzip). I can't find a way (and yes I googled it) to open it on Linux.
View 2 RepliesI was given a 1.9GB zipx file. I googled it but could not find a way to uncompress it natively on Linux (only on Windows with the latest release of Winzip).
View 1 Replies View RelatedAs the thread title says: "How can I uncompressed a .KGB file?"
View 1 Replies View RelatedI ftp upload a zip file to a linux pc, for example, somefile.zip, can I uncompress the zip file use "#tar zvf somefile.zip",
View 5 Replies View RelatedI intend on installing a version of Freebsd 8.2 under a virtual machine running on centos 5.6.The dvd image image is FREEBSD-8.2....xz. I would like to uncompress it so that I can burn it to a DVD, but have never confronted such a format. Can someone suggest which utility under Linux would be used to uncompress it?
View 4 Replies View RelatedUpon attempting to uncompress a tarball with
Code:
I have got a backfile with extension .tar.gz.enc. After fooling around I found out that it is encrypted gzipped tarball.
View 2 Replies View RelatedI have a fairly simple setup for my Network attached storage. I have a 1TB external hard drive connected to my Asus RT-N16 router through USB. I have enabled file sharing in the router, and thus the hard drive shows up on the network as a shared folder. No hassles, I can read and write to it.
Because I want to use this hard drive for networked backups, I tried to get it to mount on boot with fstab. Here is the line in my fstab file :
Code:
//192.168.1.1/SharedFolder /media/SharedFolder smbfs guest,_netdev 0 0
It mounts fine, no problems, and once again I'm able to read and write files to it. The problem occurs when I try to access the mounted drive and uncompress an archive file for the duplicity backup program. It gives me a list of messages showing that all the files inside the archive failed to uncompress because of error messages like these :
[Code]...
If I try to extract the same archive by manually going to Network > RT-N16 (Router) > Shared Folder, it extracts fine. But I need this folder to be mounted on boot because I use it very frequently. I tried changing smbfs to cifs and it still didn't work. Please not that my knowledge of samba, network storage and file systems is very basic. I tried searching the internet for more information about this but wasn't able to find much information that pertains to this particular problem. Hoping the linux pros here could give me some pointers ..
I'm running a script and the response is, missing utility no uncompress. Please install this utility. The system searched in /usr/sbin; /usr/bin/; /bin.
where I can receive a uncompress (which I guess is a .rpm) or utility that will work with my version of RHEL 5.6?
Running I386 GNU/LINUX
2.6.18-238.el5 #1 SMP
How to install ImageMagick on slackware? I can't run ./configure after I uncompress the file. I was trying to install a software and it says Imagemagick is not installed but when if I run identify -version there is some kind of version displayed but I'm not sure if it is installed.
View 3 Replies View RelatedI have a folder and its contents with the following permissions: 2750 (sguid bit). With this I ensure that a newly created file or folder inside that folder will adopt the SGUID. The problem is that if I copy files into that folder, these copied files don't adopt the SGUID. So I have to execute constantly the command: chgrp -R thegroup nameofthefolder everytime I copy or uncompress something into that folder. Is there any way to achieve this (force the SGUID even in copy and uncompress commands) automatically?
View 5 Replies View RelatedI want to search a directory recursively looking for new .rar/ .zip files. When a new file is found I want to extract the contents to another directory. To top things off would like to rename the source file as something like original.rar.extracted.
View 8 Replies View RelatedI get a SD card. Put in the SD reader. It's empty. I go to my super-important-pictures-to-a-monthly-relatory folder and select all files. Select them for MOVE. Paste them on the SD card. When the move/paste process is finished, i click on the "Eject" button on top of the SD card name. Card's ejected. I can't access the card anymore. I take out the card and put on my other computer. From 300 pictures, there are only 10 available, the remaining ones are there, but with 0bytes and unrecoveable. I panic. I go back to my main computer, my pictures are not there anymore. The pictures were on the Home folder. I panic again. I reset the computer and boot on the LiveCD. I install foremost, scalpel, photorec and about everything till my USB drive complains about being filled up. I run everything and I can't recover my files. I'm in the danger of getting fired. Things like that makes Windows sounds more appealing. When you securely remove a pendrive, things get REALLY pasted there before screwing everything up with a removal.
View 8 Replies View RelatedI know I can do find . -type f, but that includes binary file and I couldn't find a way to exclude them with find
View 4 Replies View RelatedI have installed this program ok but I am new to command lines in terminal.
I want to convert some wav files to wma files. I have the wav files currently in a folder called Test to make it easy. So I have entered the following command line:
ajpearson@ajpearson-laptop:~/Desktop/pacpl-4.0.5$ pacpl --to wma home/ajpearson/Desktop/Test and the error message I get is:
error: the following is not a file or directory: home/ajpearson/Desktop/Test
It does not matter what directory I use I get the same error. I am sure the answer is obvious - but not t me.
I have a Kingston 8gb Datatraveler that has been giving me troubles lately. For some reason after I delete files from it it still shows up as full and the files are shown in the hidden trash files. How do I get rid of these files? I can't delete them as they just show back up. Also, I tried to format the drive with gparted and it won't unmount. When I right click and select information, at the bottom it says: Unable to find mount point. Unable to read the contents of the file system. Because of this, some operations may be unavailable.
View 1 Replies View RelatedIs there a cc cleaner equivalent for linux to clear out tmp, internet files, etc.?
View 9 Replies View RelatedI am using the diff command with the -r option, to compare a large number of files and files in subdirectories. My main interest is to find out which files have been changed, and not what the actual changes are, and since a lot of files has been changed, it would be a lot easier to view the file names only. Is there and option for diff that might do this, or does there exist a similar tool/command that could do the job?
View 1 Replies View RelatedHow can you create a script to move or copy files from a main directory into multiple directories below the main directory.
View 7 Replies View Related1. Every Sunday2. Find all files older than 1 day3. Gzip these file4. Tar up the gzipped files into one tar file.5. Name the tarball with a date stamp indicating what day it was created, so we know that week's files are in the file
View 3 Replies View RelatedI have bought an external usb hard drive on which I back up my three computers every once in a while.Space will quickly be used up.I can't find that little bit of research that I need yesterday.Here is what I would like to find:An application that eliminates doubles in identical files and renames files that have changed by appending the last saved date yyyymmdd to the file name.Does such an application already exist?
View 6 Replies View RelatedUntil now i haven't had to dabble with bash scripts.
I have a program that reads in data files. These are named datafile01_R, datafile01_G, datafile01_B, they then increment, so datafile02_R etc i have about 600 of these. the program reads in 3 data sets at a time from each run, so files_01 r, g, and b.
The program then does its magic, and outputs about 40 different files, depending on the file, they gone to folders named R, G, B, psa, or tracking.
The program itself has configuration files to say where the files should gone when analyzed, there is also the config files that reads in the data sets.
At the moment i have to run one set of data, then go in and manually change the input file location, and run again. But, doing this, even though a different data set, the new set overwrites the old set in one of the output folders. So i need a way to increment the output filenames after they are written and before the program is run again with the new data set.
I've discovered that Dolphin seems to lose random files when copying many large folders.
I first noticed this a few months ago when I tried to copy my music library from one folder to another on the same HDD. It consisted of around 600 folders and 6500 files. During the copy there were no errors but after the copy I found that some of the newly copied folders were missing files. I put it down to human error or a glitch.
Yesterday I tried to copy 13 folders containing rips of some of my DVDs. Each folder basically had one film of either 700MB or 1.4GB. Again no errors showed up during the copy but I found 3 of the newly copied folders were empty.
It's not so critical with music or films but I can't afford to lose work data like this.
Has anyone experienced or seen a similar problem with Dolphin? I'm going to have to do some more extensive testing but this is not good.
The first time I noticed the problem I was running KDE4.3.4 (I think) and now the latest was with KDE4.4.0.
the VBO file that i want to burn into a dvd is around 7.9 gb. i want to know if it's possible to split them in order to burn the files into two diferent dvds.
View 5 Replies View RelatedI changed to wubi last night, i decided to go that way cause i was informed it was a safe way to try giving it a try for linux and ubuntu without any special effort needed to be payed and easily deleted if i am not pleased.Thankfully i am pretty satisfied with the results, i pretty much dealt with most of the issues i have faced so far successfully and i am running it ok beside one important thing.See i am using my laptop, which has a hdd of 250GB.On my Vista there are 2 different hdd C+E by default, they separated my hdd.So while i have windows on C and E is pretty much used for my additional data(see movies, music etc) when i installed wubi i installed it on E, thought it would be better and it had more space.
Now though, while i can access threw ubuntu all my files that were located on C(vista) i cant access any of the E ones, and search file dont helps either.The "vista hdd" as ubuntu describes it, its a 250 gb disk on computer, which means it should contain both of the vista disks.But sadly thats not the case.i cant seem to be able to find them anywhere.
What would be a nice, simple command to go through all files in a directory (no sub-directories), and change all the MP4 Video files I have to MP3 audio files (keeping the original filenames except for changing the "mp4" extension to "mp3")?
The files in question were videos taken with one of those Flip cameras, but I only need the audio off of it.
I have noticed that the files and folders search in Unity, only shows up those files which have been amended (or possibly just opened) since the install.
i was wondering if there was a way i could have the search index (or something vaguely equivalent) all the files on my machine. This is especially important given that i reinstall the OS every six months on a new distribution cycle, copying all my old files across.
Without being able to see my old files the search is pretty much reduced to a recent history search.
I found this command that works great finding and replacing a simple string to another in files located in that folder and all sub-folders.
Code: find . -name '*.php' | xargs perl -pi -e 's/OldText/NewText/g'
The problem I have is that I need to replace a more complex string, like this: Old string: /mnt/stor6-wc2-dfw1/627896/982574/ New string: /mnt/stor8-wc2-dfw1/369587/302589/ There I don't know how to do it... since the / is what separates the old from the new strings, and the strings that I want to replace have / in it. Also, I would like to know how to specify under what folder replace the files, for example, I want that it search/replaces all files under /var/www/mysite/htdocs folder.
I was reading [URL] which as the following in "Warning": Warning It is not a good idea to configure InnoDB to use data files or log files on NFS volumes. Otherwise, the files might be locked by other processes and become unavailable for use by MySQL. What does that mean, and how can one configure or check to ensure the above is being followed?
View 4 Replies View RelatedI want to automate this using script.How to automate it?
File1:
s.no# 1 name:aaaaaa
city:abcd
[code]...