Ubuntu :: Compress Files Using "zip" Command Into Two Parts?
Oct 19, 2010compress files using "zip" command into two parts?
View 6 Repliescompress files using "zip" command into two parts?
View 6 RepliesI'm looking for a way to compress a large file (~10GB) into several files that wont exceed 150MB each.
Any thoughts?
I've a file with a size of 6GB. I would like to compress this file and split them into smaller files. I was also thinking in use bzip2 to compress it, because if offers a good compression rate. How can I split this file into small ones to compress it?
View 5 Replies View RelatedI am trying to compress a folder by right clicking and selecting Compress...I get the following error:
An error occurred while adding files to the archive.No such file or directory
I want the folder and its content to be compressed to a .ZIP file which is natively accessible by Windows.
in imagemagick,i have a png format picture, but it seems too big for me,i wanna compress it ...how can i do ...command line...Code:convert "png.png" -strip -units PixelsPerInch -density 96 -quality 60 "png2.png"this command can't compress the picture, the worse is: the size will be bigger...
View 4 Replies View RelatedI just read in my Linux+ resources that it is not a wise idea to compress tar files with gzip if my tape drive supports compression. In addition, the resources mention that using gzip for tar files runs the risk of data lost if an error occurs with the compression.
Does anyone compress tar files for major backups/restore policies?
When I right click on a file and click "compress" I get a .zip file that is exactly the same size as the original file.
View 2 Replies View RelatedI want to know how to compress files to zip, rar or 7z with password using KDE or Dolphin interface and no command line. I can compress without password, I have not found an option to protect with password. What do I have to install or configure?
View 1 Replies View RelatedI want to package my application for linux, I use zip/tar.gz/7z/rar to do the compress in Windows.Because I do the compress in Windows, so when I extract the package in Linux, the file permission is reset.The problem is Only the tar.gz can let me extract runable files, other format (zip/7z/rar) not (the file permission is reset to 644).So my question is how to compress my files using zip/7z/rar while reset the permission to 755.
View 2 Replies View RelatedI currently have a bash script that runs and backs up my daily files onto a separate partition using Rsync, but I thought it would be good to use the Ubuntu-one service as an ADDITIONAL backup for really important files.
How do I compress then encrypt those files, and can I add any commands that will do this to my existing bash script?
I am running Ubuntu 10.04
i have noticed that if right click on many files types for instance iso files i don't have the option to store them in compress files at all i have one iso file now that i want to compress and split to three parts so i could upload it but as i said right click on the file don't helpbecause i don't have the option there so what can i do?
View 4 Replies View Relatedi need to write a short script that will compress a specific folder that`s on the Desktop (and all it`s content) and also will encrypt it with a password that is inside the script --->meaning it wont ask for a password+verification when compressing+encrypting
View 1 Replies View RelatedIs there is anyway, with tar, zip, gzip, or any file compression type to compress without causing high CPU. In other words, limit how hard the CPU works to compress it? Of course I understand that this would cause the compression to take longer but time isn't too big of a concern.
View 3 Replies View RelatedCould someone help me find a way to rename a file to a different name containing parts of its old name?
For example:
Original file name: filename1.abc.xyz.some.other.stuff
Final file name: filename1.abc.xyz
The length of the file name is not constant. the abc.xyz is not constant but that format is (three numbers.three numbers) the .some.other.stuff is not constant and its what i want to get rid of
Today I tried to compress some folders containing backup files from last year. I right-clicked on the folders and selected compress as tar.gz. I let it work, and found that hours later, the folders were still compressing. How long is it supposed to take, anyway? I was trying to compress the two sets of backups simultaneously; together they're around 1.5 GB. They have many subdirectories.
View 4 Replies View RelatedI'm creating a script that creates files from svn checkout and compress them using tar.gz the script gets the repository name from command line argument i need to capture a number from the last line of the output and create a file name from it.
The svn returns output of all the file names from the repository and in the end it says: revision number xxxxx. i need to get this number and then rename the tar.gz to it. how do i save the output to a variable and get this number.
This is the script:
Code:
I download a lot of files that have been compressed into separate pieces by winrar, so that a 4gb file will be shown as dozen or so 93mb peices. In windows, using winrar, extracting them was rather easy but in Kubuntu whatever default program is being used throws a fit. In a perfect world I would want to get a open source alternative from synaptic, alternatively anything I can download in RPM I can eventually figure out. I want to avoid trying to get winrar working under wine.
View 13 Replies View RelatedI used split -b 32m "file.bz2" "file.bz2.part-" to split a file and it created more than 50 parts. From googling, the way I found to reassemble the parts is to cat file.bz2.part-aa file.bz2.part-ab > file.bz2, while enumerating all the 50+ parts. Is there an easier way to reassemble the parts wherein I no longer need to list all those parts explicitly?
I'm using Fedora 12.
I am facing a problem while splitting a text file, I need to split a file into some parts and each split file should have 2000 lines, when I do it through "split" command the mother file is kept intact but as per my requirement I need to cut mother file into some parts thus it should not be kept intact.
Example:
file size
motherfile.txt 5000 lines.
after split
motherfile.txt 2000 lines.
childfile1.txt 2000 lines.
childfile2.txt 1000 lines.
Today encoders are getting smarter. They can compress Blu ray similar quality in 700MB. It seems header of video file contain info about frame rate, audio/video encoder etc. which can't be guessed. In MPEG audio , every part of file is independently playable. If a movie is binary split into 6 parts & I don't have the first part then it is unplayable.
Code:
example
ls
-rwxrwxrwx 1 root root 280M 2010-12-07 20:23 irn2-cd1.mkv
-rwxrwxrwx 1 root root 50M 2011-05-26 13:09 last-50M-cd2
-rwxrwxrwx 1 root root 50M 2011-05-26 13:44 first-50M-cd1
file *
first-50M-cd1: Matroska data
last-50M-cd2: data
irn2-cd1.mkv: Matroska data
I am using the diff command with the -r option, to compare a large number of files and files in subdirectories. My main interest is to find out which files have been changed, and not what the actual changes are, and since a lot of files has been changed, it would be a lot easier to view the file names only. Is there and option for diff that might do this, or does there exist a similar tool/command that could do the job?
View 1 Replies View RelatedHow can i Compress a file in ubuntu 10.04??
View 5 Replies View RelatedIs there a method at the command line to copy files from one location to another and retain the source files group and user?I'm migrating some MySQL files from one machine to another.I want to back-up the original files in the directory presently. They have owner:group of mysql, some have owner:group root:mysql and so on. To copy them under cli or Nautilus everything changes to root for I execute sudo cp or gksudo nautilus and copy via gui.
Since it is MySQL data I could simply do a dump of the database and restore it on the other machine. But there's about 20 db's and I want to do this via a copy for it will be faster - at least that is what I think.
I have a large list of folders that I want to compress using tar.the script I want to right is to travers all the folders and for each I would tar the contents to a tar file with the same name of that folder. and the it should be stored inside the same forlder.
View 6 Replies View RelatedI have a compressed text file. The method of compress is unknown.I can see the file contents by using Midnight Commander without a problem but I would like to view the file just with cat. So I am trying to uncompress the file with unzip or gunzip but it does not work.How to check the method the file is compressed with? Is any way to find it with Midnight Commander?
View 4 Replies View RelatedI want a huge file of .jpeg files. How can I can compress these files to save my space? Is there any s/w for this??
View 2 Replies View RelatedI found this command that works great finding and replacing a simple string to another in files located in that folder and all sub-folders.
Code: find . -name '*.php' | xargs perl -pi -e 's/OldText/NewText/g'
The problem I have is that I need to replace a more complex string, like this: Old string: /mnt/stor6-wc2-dfw1/627896/982574/ New string: /mnt/stor8-wc2-dfw1/369587/302589/ There I don't know how to do it... since the / is what separates the old from the new strings, and the strings that I want to replace have / in it. Also, I would like to know how to specify under what folder replace the files, for example, I want that it search/replaces all files under /var/www/mysite/htdocs folder.
i am downloaded some e-books in the format .rar. when i am extracting them i am getting error as There is no command installed for RAR archive files. Do you want to search for a command to open this file?
View 3 Replies View RelatedI have a very dumb questions regarding run-parts utility. I've created few shell scripts, chmod them to be executable, placed them into one particular directory and tried to run them using run-parts utility like this
run-parts dir_name However it doesn't seem to be working or I might just be missing out something. I have even tried different options like --test or --list - NONE of them produced any output. If I run each of the script manually they all work just fine. Is this is a known bug or somethng ?? or am I really missing out something...how are the /etc/crontab entries then being executed when the run-parts doesn't seem to be working on such a simple case ? ps: I noticed one interesting thing though...when I run run-parts like this: run-parts --list --regex 'some_regexp_patern' dir_name -> the output actually produces a list of the files whose file names match the given regexp pattern .
How do I cut jpg images in two parts?
If I want to e.g. slice an jpg vertically at so and so many pixels from the bottom.