opensuse v11.2, linux 2.6.31.12-0.1-desktop x86_64
ZIP v2.32
I wish to exclude some files from a zip archive. On other OSes to exclude an entire directory I would use the "-x" option like so:
Code:
zip -r archive-name * -x dir1/* Simple. And just add "-x"'s as needed (or use an exclusion file).
Not how it works here, it would seem. AFAICT all "-x" options are ignored. (The entries in an exclusion file also.) For instance, "-x diy/mplayer/*" should ignore everything in the <diy/mplayer> directory. It does not. I have tried fully qualified paths as well; no joy.
i have created on folder in my server to upload some regular states. I want that user can modify or upload already stored files. but, should not upload any unwanted files orfolders.for that i want to use "rm" command as auto scheduler (putting this in cron tab.so that all files will be removed except some required files / folders for which this upload facility is activated. users are using secure-shell for uploading data.
I'm trying to find all java files in bash that contains the method "assign()".I would like to retrieve the same list except without the Test* files. How can I do that?
i am downloaded some e-books in the format .rar. when i am extracting them i am getting error as There is no command installed for RAR archive files. Do you want to search for a command to open this file?
I was thinking of migrating my apt-mirror repository to the recommended ftp scrips: [URL] .....
I pre populated my pool with already downloaded files, and setup the scripts.
However, if I run the bin/ftpsync, and monitor rsync with lsof -p, I can see that it is still downloading files from oldstable (wheezy) despite exclude options.
I'm guessing it's a configuration error, but I can't seem to figure it out. Any thoughts? My etc/ftpsync.conf is as follows:
Actually, I don't think it works like I thought it did. A few guides I found listed the exclude options, but the sample config file has this:
Code: Select all## If you do want to exclude files from the mirror run, put --exclude statements here. ## See rsync(1) for the exact syntax, these are passed to rsync as written here. ## DO NOT TRY TO EXCLUDE ARCHITECTURES OR SUITES WITH THIS, IT WILL NOT WORK! #EXCLUDE=""
So it looks like it doesn't exclude the suites at all.
In reading the rsync man page and browsing a lot of websites, I ended up a bit confused, or maybe it was just too much eggnog. Anyway, to exclude a directory "videos" with everything in it, which is /home/user1/camera/videos and I'm rsyncing the whole user1 directory to an external drive
I have the following command which finds all files that have changed in the last day and lists them. How can I exclude hidden files like .bash_history?
I'd like to ask about archive mounter feature, can I mount zip file with read write mode? can gvfsd-archive do that?, or I must use fuse-zip to mount it? If I must use fuse-zip, how I wrap it so I can use it via nautilus or via gvfs-fuse-daemon
I have a ton of music that I would like to upload to an on-line storage site. Ideally, I would like to keep the file organization structure intact, but am limited to 50mb file size.
Is there a way to compress this whole directory without going through and compressing every folder individually, while maintaining less than a 50mb file size?
I have a .tar archive of music that I would like to extract specific files / folders from. I've been looking at man tar but I don't seem to be able to find any information related to my quest, although someone will likely point out that it's right there in front of my nose
Anyhow, the file is called music_archive.tar and I would like to extract the folder /music/P/Pink Floyd/The Wall/. Anyone know how to do this
If I understand everything right I can use simple cd to create an install of my system then config everything and create an archive of /etc,/var,/usr and that give me a backup of my changed files. but what I am not able to do is create a script to replace the files with the archived version of the config files automatically. I need it to be ran as a post intall script or to configure the conf files as part of the install. what I configure on every reinstall. apache,proftpd,mysql,php,and wordpress. Since I am still learning I reinstall a lot and standard backups have not been working for me. But I can put the whole install including backup on one cd. Is there a way replace only the change files from the archive. or getting a list of just the modafied files on my system.
I have a RAR archive, split into 40 sections, containing a folder with a total of 2GB of data. When I Open the Archive using "File Roller 2.30.1.1" and drag the archived folder to destination folder, it says that its uncompressing, but after its uncompressed, there is nothing there, in the destination folder. Where was the 2GB extracted stuff gone?
surprisingly come across this file that the ubuntu archive manager is simply unable to unpack.these are some articles that i am trying to download.strangely i have never heard of the format and the page asks me to download "stuffit expander" to allow me to view the pdf files. however when i download the file saving it on the desktop, and try to open it, archive manager is simply unable to do so!this is the title of the downloaded file -
I have just entered the world of openSUSE after a chain of other linux distros over the last few years. Btw I love it, its already my favourite. It is useless for most people who are behind a router, because any transfers between same-network computers are subject to the same rules = slow and unnecessary.I used #256461 - wondershaper: Option to limit range of rules? - Debian Bug report logs for these mods.
Suggested /usr/sbin/wshaper.htb
Code:
#!/bin/bash # # Wonder Shaper # please read the README before filling out these values
[code]....
Should this be filed as a bug report? It seems like an obvious feature to me.
I'm trying to create backup/archive my Ubuntu 10.04 system files (so I can restore it in case my system get corrupted). More specifically, I'm trying to zip the important files in my root directory not including my home directory (which includes my documents which I backup separately/more frequently) to an external hard drive attached via USB (called 'My Book').
Since File Roller didn't give me quite the level of control I was looking for, I created a script that I could execute to backup and archive regularly. Here's a snippet: cd /media/"My Book"/"Linux Backups" NOW=$(date +"%b-%d-%y") LOGFILE=Backup_Root_FileSystem-$NOW.log sudo zip -r -T -v Backup_Root_FileSystem-$NOW / -x /media/'My Book'* /media* /proc* /sys* /mnt* /dev* /cdrom* /home* /'lost+found'* | tee -a $LOGFILE
When trying to create a new compressed/archive file in Gnome Commander (GM) the file is created but the selected files are not added. I can open the new (empty) archive file and then add files to be compressed. I have tried using several different formats (zip, tar.bz and others) with the same results. The "file roller" is shown as a plugin but has no configuration other than the compressed file type.
AKA "zipping on the fly .. the slow-as-molasses way." The list includes full pathnames to each file, and they're all in subfolders of the same parent folder (which, unfortunately, is not the root folder of the drive or system on which the files reside). A cleaned-up and radio-ready portion of the list looks like
What I'd like to be able to do is zip all the files in the list into a single archive, to avoid the step of having to copy them to the same location (presumably another folder on the HD) and then zip that folder. I'm more inclined to make provisions about extracting to a single folder at some other time. Is this possible in BASH, or would I have to consider a faster, more robust scripting language such as python or perl?
I have to administer a few mail servers, a mail log server, 4 nameservers and a web server -all running on Centos 5 server distributions. Now I have a task: to avoid accidental crashes on the production servers while installing updates, my boss asked me to do clones (these clones will all be VMware virtual machines) of the servers (EXCLUDING the actual e-mails and mail log contents) and then to run those clones on VMWare Server. This way, first I will install and test updates on the clones and - if they will be running without crashes - I will apply the updates on the real production servers themselves. I have already installed VMWare Server 2.0 I have a few questions:
- How do I build the virtual machines to exclude the actual mail files and mail logs? Can I use VMware Converter for this purpose, or do I have to use another program? - How do I actually do this cloning? Is there a tutorial on how to do this?
networkmanager kde installed I have to add the configuration of a VPN using a archive .ovpnBut in knetworkmanager vpn tab, not see the option to import. Only displays "add, edit and delete".How I can add this option?
I am new to the world of linux and when attempting to verify a tar archive I am displayed the following error. When running the command tar cvfW archivename.tar filename directoryname does not yield any errors.
All operation result in a seperate window displaying the progress of the compression but with an error.
Error 127, cannot execute requested operation.
Then when the progress bar reaches full it resets and then resets again continuing the same loop with it becoming slower with every repetition. There is no information posted in the report log of the window. Except that the task has started.
I'm trying to simply archive a file and password protect it. This shouldent be such a difficult task.
P7zip also gives me its own set of errors when archiving.
Right now I use "rsync -a" to sync my zimbra directory with another directory as a backup. I'd like to archive that backup (zip....rar....etc etc) in such a way that it preserves the same information as rsync -a would (symlinks, permissions.... Etc etc). Does anyone know of such an archive method? My goal is to have it so that if I uncompress the file, the result should match exactly what my zimbra was at the time of backup.
I came across this feature of the "less" command today. It appears not to be well-documented so I'll share it here. I found that "less" can automatically uncompress compressed files - no need to run the first command, the second does the same job:
Code:
gunzip -c textfile.gz | less less textfile.gz
Even better, "less" can read and list the contents of some archive files as well, including Tar files compressed using Gzip and Bzip2 as well as Zip files. So now when I want to list the contents of an archive,for example to determine whether the contents will be extracted to an absolute path or relative to the working directory, the following work
Code:
less archive.tar.gz less archive.tgz less archive.tar.bz2 less archive.zip
Much easier than remembering the various options that the Tar, Gzip, Bzip2 and Zip use. The best one for me is with RPM archives:
Code:
less some.rpm
The above outputs the Information, Change Log and verbose File List of the RPM (paging the output, of course!) The following commands (using short and long options) give the equivalent output using the "rpm" command:
Code:
rpm -qilvp --changelog some.rpm | less rpm --query --info --list --verbose --package --changelog some.rpm | less
"less" does not recognise the initramfs file produced by "dracut" (which is a compressed "cpio" archive), but I suppose you can't have everything ...
According to the man page, rsync should copy "special" and "device" files if the corresponding options are set. Also the man page says that the -a option (--archive) includes the -D option, which is "same as --devices and --specials". BUT We use the -a option, and still get the following in our report:
I'm new to Ubuntu, and everytime i've tried to download a program like iTunes, the "Archive Manager" comes up and says "An Error Has Occured While Loading the Archive". how to fix this or download programs ?