General :: Get Error "does Not Look Like A Tar Archive" When Attempt To Verify An Archive Using Tar?
Jul 13, 2011
I am new to the world of linux and when attempting to verify a tar archive I am displayed the following error. When running the command tar cvfW archivename.tar filename directoryname does not yield any errors.
I'd like to ask about archive mounter feature, can I mount zip file with read write mode? can gvfsd-archive do that?, or I must use fuse-zip to mount it? If I must use fuse-zip, how I wrap it so I can use it via nautilus or via gvfs-fuse-daemon
I'm new to Ubuntu, and everytime i've tried to download a program like iTunes, the "Archive Manager" comes up and says "An Error Has Occured While Loading the Archive". how to fix this or download programs ?
7-Zip 9.04 beta Copyright (c) 1999-2009 Igor Pavlov 2009-05-30 p7zip Version 9.04 (locale=en_US.UTF-8,Utf16=on,HugeFiles=on,4 CPUs)
Error: there is no such archive
this is what i got when i tried to download cinema 4d on my computer 40 mins. ago, is something im not doing that makes this happen because this happens more often than not, i go through a lengthy download and get errors, im sorry ive only been on here for 2 weeks tops and already bothered you all 4 times.
All operation result in a seperate window displaying the progress of the compression but with an error.
Error 127, cannot execute requested operation.
Then when the progress bar reaches full it resets and then resets again continuing the same loop with it becoming slower with every repetition. There is no information posted in the report log of the window. Except that the task has started.
I'm trying to simply archive a file and password protect it. This shouldent be such a difficult task.
P7zip also gives me its own set of errors when archiving.
I had been running a dual boot windows XP & Ubuntu. Graboid worked in both systems. I have now got rid of windows, upgraded to 10.10 I installed wine, winetricks & dotnet20, but when I click on my downloaded Graboid.exe file I get the following error: Command line output code...
I'm trying to install Adobe Reader 8 and the command #dpkg -1 AdobeReader_enu-8.1.3-1.i386.deb comes up with the message error processing AdobeReader Cannot access archive. No such file or directory. Error was encountered while processing.
I am stuck trying to write a script that does the following :
1. loops through all subdirectories of a given directory
2. for each found subdirectory, first create an archive carrying the same name as the subdir itself
3. then moves the actual subdirectory to a different path
Here is what I have so far: my base dir is /home/bob/Bureau, and it contains two sub dirs, "florissant 86 a" and "saule 84" I would like to create one archive for each subdir in /media/public/atelierPhotoArchive and then move the folder entirely to /tmp/photo.write
Everything goes well until I have to either tar or loop through the file with spaces in names
This last statement outputs different lines after each space in my lst file
Question 1: Is it possible to make it output once for each line in the file ?
Question 2: Is it possible to do some exec tar in the find command? I had difficulties extracting the "short" name for the archive ("saule 84") without /home/bob/Bureau, it is possible only with the printf %f, but how in the world can I get this value in to the exec option ?
I want to untar a package from one directory to another directory, directly from the command prompt. I want to untar the joomla package into the htdocs directory of xampp. How to do that directly from command prompt ? The reason i am asking this is if I try the "drag and drop" way, it won't be possible as xampp is stored in /opt directory and without super-user authentication nothing can be saved into it. You can argue saying that why did I untar xampp initially in su mode ,but that had to be done so that apache doesnt give me any start-up problems.
I'm trying to use tar to tar files before transfer, so I can keep the entire file path rather than losing it along the way. However, when I try to tar an empty folder, it tells me that it is cowardly refusing to create an empty archive. I want to keep the empty folder on the other end, but don't want to put anything else into the archive to make it non-empty. Is there any way to do this?
I need files to be <= 5GB to put on S3. Right now I have an ugly tar / gzip / cut before upload, then cat / zcat / tar on download, but it's really ugly - and nearly every archiver should support archive splitting right? What's the best way?
I'm searching for a tutorial on how to convert my videos (AVI, WMV and MKV formats) to a format that offers both good quality and small size.The audio part should be 128 kbit/s mp3 lame.I'm using Fedora 14 with an rpmfusion repo. It would be especially nice to be able to batch-convert the AVIs.
I have been playing around with the tar command and I know this is how to use it. Code: tar -cf [filename] [directory] But what I want to make an archive from the current directory I thought just to not enter a directory but that doesn't work. I get an error about creating a empty archive so how to do I make it so how do I tell it to do the current directory?
The command tar -xvf wpa_tables.tar gives this error
tar: ./xai-0/334Regency: Cannot open: Invalid or incomplete multibyte or wide character tar: Exiting with failure status due to previous errors
Obviously the backslash is giving tar some problem, but I've been all over the docs and can't figure out how to either make it skip this file or interpret the character literally. Here's my command history to show some of the options I've tried which don't work.
tar --no-wildcards -xkvf wpa_tables.tar tar --exclude ./xai-0/\334Regency -xkvf wpa_tables.tar tar --exclude "./xai-0/\334Regency" -xkvf wpa_tables.tar tar --ignore-command-error -xkvf wpa_tables.tar tar --ignore-failed-read -xkvf wpa_tables.tar tar --transform 's/\/slash/g' -xvf wpa_tables.tar ./xai-0/\334Regency
I'm totally out of ideas at this point and would welcome some input from more experienced members.
Ubuntu 10.04. As part of my nightly backup script I archive my home directory with the following command tar -cvpzf /quitelarge/_mirror/mirror1/home-ken.gz /home/ken 2>> /quitelarge/_mirror/tar-error.log
It seems to work fine and I have recovered files from the archive on occasion. Actually I keep 7 rolling daily backups and a monthly burn to DVD. I had an sftp connection made by Nautilus to my server. Ubuntu for whatever reason places an icon on the desktop showing the connection. When I ran the script it decided to archive everything on my server - all 1.4 TB. I caught the problem when home-ken.gz was about 5 GB. I stopped the process, closed the sftp connection, rolled back the backups and tried again. This time I got a file of the expected size - about 45 MB.
In the error log I did find that the tar process was trying to suck the entire contents of the server into the archive file. tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/asound/ICH5/pcm0c/sub0: file changed as we read it tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/asound/ICH5/pcm0c: file changed as we read it tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/asound/ICH5: file changed as we read it tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/asound: file changed as we read it tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/scsi: file changed as we read it tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/acpi/event: Cannot open: Permission denied tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/acpi/fadt: Cannot open: Permission denied tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/acpi/dsdt: Cannot open: Permission denied tar: /home/ken/.gvfs/sftp for ken on taylor10/proc/irq/21/smp_affinity: Cannot open: Permission denied
Is there an option I can place on the tar command to tell it NOT to follow the ssh connection which is sitting on my desktop? The closest thing I see in the documentation is -h which tells tar to "follow symlinks; archive and dump the files they point to." I am NOT specifying -h so if the ssh connection is treated as a symlink by tar I would still not expect the remote contents to be tarred.