Ubuntu :: RAR Archive - Extracted Files Not In Destination Folder
Sep 14, 2010
I have a RAR archive, split into 40 sections, containing a folder with a total of 2GB of data. When I Open the Archive using "File Roller 18.104.22.168" and drag the archived folder to destination folder, it says that its uncompressing, but after its uncompressed, there is nothing there, in the destination folder. Where was the 2GB extracted stuff gone?
trying to run a script every 30 minutes that checks a directory recursively and extracts any archives that have yet to be extracted.
I tried the unrar o- option which does not overwrite, but it still goes through the motions of extracting the files, then just skips the overwrite so its wasting cycles and HDD life (especially if running every 15-30 minutes).
What are my options?
I really just want:
if (archive has already been extracted) then (skip extraction) else (extract archive)
is there any simple way to accomplish this?
My formatting kept getting jumbled into one big paragraph so I tossed it in a quote tag =/
I cannot change directory to a more than three folder tree destination folder from ~ in terminal. I've checked everything. No Typos or misspell. The destination folder was recognized by "ls" command but when I went to it, the terminal said, "no such file or directory."
I'm trying to do the simplest thing. Compile XMMS. I navigated to the folder containing the extracted source and ran ./configure. That failed, so I remembered to install the build-essential package.
After that, I continued and got this output:
It said I didn't have glib installed, whatever that is. I did "apt-get install glib" and it told me it couldn't find the package called "glib". I thought, "Yeah, that seems a little too broad of a package name".
So then I went to Synaptic and did a search for "libglib" and noticed I already have libglib2.0-0 installed, as well as libglib2.0-data and libglib-perl.
How can I get this to to work correctly? I know the process should be ./configure, make, and then make-all, but I get stuck on the ./configure step.
I got the etpro 3.2.6 zip file and extracted it but there are no files at all that can be run in linux inside the folder even though it was supposed to be a linux download. Is there something I am missing and is there an easier way to install 3.2.6?
I've extracted a few files through the archive manager through Ark (as well as through the command line) and some folders/files that were originally in Japanese are not properly displayed in the terminal or file browser. I can't delete them nor open them; I am informed that this file does not exist.I am running Kubuntu 11.04. I had a similar problem in Ubuntu 10.04 in that the text was not displayed properly, but the files were readable and writable.
I've written the script below to merge only .txt files that exist in one directory into one huge .txt file and ignore other files with other extensions. now the result is one huge .txt file with all the contents of other .txt files
how can i add the File Name as a comment before each file?
I have extracted the files from a bootable CD to /dev/sdc2/bootable/winpwrec with the hope of booting from the hard drive. I have read the grub.cfg file and a couple of guides here and I'm still not sure what I'm doing. Is there any chance someone can tell me what I need to add to the grub.cfg file to add this to the boot menu? I am pretty well-versed in computers but I'm new to Linux and I'm still trying to get acquainted with the OS, terminology, and the boot loader. I'm using Ubuntu 9.10, if that matters.
files in the folder: BOOT.CAT;1 ISOLINUX.BIN;1 SCSI.CGZ;1 BOOT.MSG;1 ISOLINUX.CFG;1 SYSLINUX.CFG;1 VMLINUZ.;1 INITRD.CGZ;1 README.TXT;1 SYSLINUX.EXE;1 isolinux.cfg Code: #serial 0 19200
When K3B is creating an .iso image to file and fails to read a sector even at the very end of a DVD it issues you a warning and removes the file from the destination folder, regardless of what choice you made in the "remove file when completed" dialog.
I want it to leave the incomplete .iso image there so I can play with it, any suggestions cause I've searched scratched and dissected, trying to do my homework but doggone if I can figure it out,
Many years ago, I converted a portion of my files to an arbitrary format with a specific extension. i no longer desire to have them in this format and i would like begin the process of replacing them because conversion is not an appropriate solution. unfortunately, they are mixed in separate folders of the same root folder with files in my current format of a different extension. I feel it would make this process easier if I were to move every folder that contained a file with the undesired format to a separate root folder. The files are stored on a Linux server and shared via samba. How can I do this with a couple of commands or a script? I am open to other suggestions as well. I want to avoid time spent editing text files. Ultimately, I'd like a command that produced a list of full paths for folders, sorted by the number of levels would be a nice touch. A list of all of the files is clearly not what I'm looking for.
i am downloaded some e-books in the format .rar. when i am extracting them i am getting error as There is no command installed for RAR archive files. Do you want to search for a command to open this file?
I have two non-system drives with archive files. Each drive is formatted with one primary partition only, occupying all the drive space. In each drive there will be a number of directories with files in them, like this:
I have recently purchased an external hard drive in order to backup my home partition. In my PC I have a "1.5T" drive with several partitions on it, containing OSes and the home partition. The home partition is 1.3T according to df, the external drive contains one partition that spans the entire disk,df reports it as 1.4T in size. Both partitions are ext3. When I use rsync to copy files from the home partition to the external partition, the external disk becomes full, despite the destination - supposedly - being larger than the source. I don't understand why copying files from one partition to a slightly bigger partition should need more space than on the source partition. Does anyone know what is happening ?
Details : I created the partition on the external drive with gparted; gparted reported it the already have several gigabytes in used space immediately after the partitions creation - I thought at the time that this must be normal. The home partition contains many files of all sorts, including lots of big audio and video files. If you are wondering, for all my important files this external disk is only secondary backup, as they are also backed up to the "internet".
These are the mount points :
/mnt/tmp/ : home partition, /dev/sdb6 /mnt/external/ : external partition, /dev/sdc1
I used rsync to copy the files, I know there are more efficient ways to do this, but I wanted to use the same command that I will subsequently run to sync the backup.
Next I tried adding the --sparse switch, as I was wondering if the problem may come form sparse files. I don't know however if rsync would go back and shrink the sparse file by just adding the switch and executing the command. I also added --one-file-system, for good measure. Here is what I ran next :
rsync: writefd_unbuffered failed to write 4 bytes to socket [sender]: Broken pipe (32) rsync: write failed on "abcd.avi": No space left on device (28) rsync error: error in file IO (code 11) at receiver.c(302) [receiver=3.0.6]
Looking at the destination after a partial copy seems to indicate that the problem is not symbolic links being "expanded". I have not checked the source filesystem for sparse files, nor the destination to see if these files could be larger there, as this does not seem trivial.
Here is some additional info :
$ df /mnt/tmp/ Filesystem 1K-blocks Used Available Use% Mounted on /dev/sdb6 1415342836 1414173740 369096 100% /mnt/tmp
We have an rsync cron job set up to mirror all the files in a "..dashtdocsdocs" folder to the same folder on another server. It copies all the files over correctly and deletes any files in the "docs" directory that aren't in the sending directory, but it also deletes any files we put in the target directory's parent folder (..dashtdocs or other subfolders like ..dashtdocsimages) even though they've been excluded in the .rsync-filter file.
So for example server A has ..dashtdocsdocs and ..dashtdocsimages. Server B has ..dashtdocsdocs but if I manually copy the images folder over to ..dashtdocsimages, the images folder gets deleted from the target directory every time rsync runs.
I'd like to keep just the docs directory synched and update other folders manually, but they keep getting deleted. It looks to me like it's running a delete-excluded option, but that option wasn't used.
Today i am trying to learn how to use sed. I set up a testing folder with the following files:
AAb.lol AAc.lol AAx.lol test.sh
My goal is to create a script (test.sh) which renames all the files to their original name without AA. I want to end up with this:
b.lol c.lol x.lol test.sh
sed seemed to be the perfect tool so i went ahead and created a script which i think should clear the job.
mv: missing destination file operand after `$i' From that 2nd line i can tell that $NewName is just empty. I also read something about sed needing the -e option for scripting purposes but i just don't understand it.
i want to copy one or more files or directory from one drive to multiple drive simultaneously. It like a cloning a disk. But i dont like clone entire disk. i want to copy/clone only certine files or folder. if any one can know how to copy one source to multiple destination simultaneously.
I'm currently learning to use rsync to backup my music collection. I have a Firefox tab open to the rsync manual page(s) and have been reading man rsync and running experimental rsync operations.I've been doing this for the last 3-4 hours. I've used rsync for this purpose in the past with disastrous results. What was and is once again (due to a month and a half of file pruning) a 9000 file music collection had mysteriously grown to over 25,000 music files and 80GB of data! This was likely due to the fact that I didn't really know what I was doing with rsync and had never spent too much time learning about all the parameters, what their functions are and how they may relate to my goal.Here are the particulars:
* Source drive is a 500GB disk, /media/sata500/music/.
* Destination drive is a 250GB USB disk, /media/FreeAgent/music, connected to the same computer that houses the 500GB disk.
* I want to copy or backup files from /media/sata500/music to /media/FreeAgent/music.
* I do not want to create ANY duplicates of files that exist.
* I only want to add files to the destination drive if they are new on the source drive, like if I rip a CD and add the contents to the source. I want them copied over next time I run rsync.
Here's the rsync command in it's most recently used form, and probably very immature at this point.
This appears to have copied all files and folders and I'm satisfied that my goal has been met with some success. To convince myself of this I ran the command and then once it was complete I added 2 new songs putting them in their respective folders on the source drive and ran the same command again. The resulting output was
Two files transferred. Exactly what I want.Both folders now house 20,931 files and use 40.6GB. Identical as far as I can tell.What I'm concerned about are time stamps and play count data, etc. Anything that changes the original file. I don't want this data to cause a file to be transferred as I'm afraid that the new file will be created along side the old file of the same name thereby starting this whole music collection expansion thing all over again. I've invested a lot of time and effort to get it pruned down to where there are virtually no duplicates and albums are correct in that they contain the proper songs in the proper order.
I'd like to ask about archive mounter feature, can I mount zip file with read write mode? can gvfsd-archive do that?, or I must use fuse-zip to mount it? If I must use fuse-zip, how I wrap it so I can use it via nautilus or via gvfs-fuse-daemon
I recently replaced my windows fileserver with one running Ubuntu. One thing I've noticed (which is a annoying) is that when I copy files between two samba shares from my windows machine, it copies the file through my PC to the new destination. On windows shares it just did some sort of local copy (ie it took about 2 seconds) rather than 3-4 minutes. Is this the normal behaviour, is there any way around it on Linux
surprisingly come across this file that the ubuntu archive manager is simply unable to unpack.these are some articles that i am trying to download.strangely i have never heard of the format and the page asks me to download "stuffit expander" to allow me to view the pdf files. however when i download the file saving it on the desktop, and try to open it, archive manager is simply unable to do so!this is the title of the downloaded file -
opensuse v11.2, linux 22.214.171.124-0.1-desktop x86_64 ZIP v2.32
I wish to exclude some files from a zip archive. On other OSes to exclude an entire directory I would use the "-x" option like so:
Code: zip -r archive-name * -x dir1/* Simple. And just add "-x"'s as needed (or use an exclusion file).
Not how it works here, it would seem. AFAICT all "-x" options are ignored. (The entries in an exclusion file also.) For instance, "-x diy/mplayer/*" should ignore everything in the <diy/mplayer> directory. It does not. I have tried fully qualified paths as well; no joy.