General :: Compress A Large File Into Smaller Parts?

Aug 18, 2011

I'm looking for a way to compress a large file (~10GB) into several files that wont exceed 150MB each.

Any thoughts?

View 2 Replies


ADVERTISEMENT

General :: Split Huge File Into Small Parts And Compress Them?

May 30, 2011

I've a file with a size of 6GB. I would like to compress this file and split them into smaller files. I was also thinking in use bzip2 to compress it, because if offers a good compression rate. How can I split this file into small ones to compress it?

View 5 Replies View Related

General :: Large Gtk+ Controls - Making Them Smaller

Sep 16, 2010

One thing that I didnt like about gnome is the huge gtk+ interface and controls. Everything is so large.
Is it by design? Or I can make the overall interface smaller by some tweaking? If so then how? This question arose while I was using eclipse in fedora gnome.

View 1 Replies View Related

Fedora :: Compress A Folder By Right Clicking And Selecting Compress - No Such File Or Directory

Feb 18, 2011

I am trying to compress a folder by right clicking and selecting Compress...I get the following error:

An error occurred while adding files to the archive.No such file or directory

I want the folder and its content to be compressed to a .ZIP file which is natively accessible by Windows.

View 6 Replies View Related

Ubuntu :: Compress Files Using "zip" Command Into Two Parts?

Oct 19, 2010

compress files using "zip" command into two parts?

View 6 Replies View Related

Ubuntu Installation :: Igrating OS Install From Large HD To Smaller One

Apr 2, 2010

I need to use the drive which currently hosts an Ubuntu 9.10 server install. The Os install was the default Ubuntu partitioning. It created sda1, sda2 and sda5, Linux, Extended and Linux/Swap partitions, respectively. The start and end blocks for sda2 & 5 are the same. I am trying to figure out which method of migration would be the best. I have used ddrescue to clone hard drives and am not sure how this would work when migrating to a smaller HD (Moving from a 74GB to a 36GB). Would it be a good idea to shrink the partition with Gparted or something? The OS is only about 3-4GB.

The other way I could migrate is to use the tar backup I made. It should have everything needed as I used the notes from the Ubuntu Backup How To page. I just don't know how this works when installing to a new hard drive. Would I need to create or edit the MBR of the new hard drive? Can a tar backup capture any of the MBR info?I guess it shouldn't make much of a difference as to which method I use so I am open to suggestions as to which to try and any suggestions as to things to make sure to do.

View 3 Replies View Related

OpenSUSE Install :: Desktop Display Size Too Large Want Smaller As Default

Dec 28, 2009

I am using OS 11.0 Every time I boot my laptop (dell inspiron 9300 - ati video M300). I get the desktop display as 1920 X 1200. This is too large for my default. I use KRandRTray to resize back to 1024 X 768. How can I set 1024 X 768 as the default but still have the option to go to 1920 X 1200?

View 3 Replies View Related

General :: Compress The Size Of ARM9 Executable File

Apr 28, 2011

i am using classpath-0.98,jamvm-1.5.4 and arm9 cortex processor. so my question is after install classpath and jamvm on arm9 , i am getting around 30MB FAT file. so tell me some tips how to reduce the size of FAT file as much as possible.

View 1 Replies View Related

General :: Write A Script To Compress Any File With 10 MB Size To Tar.gz?

Jun 7, 2010

write a script to compress any file with 10 MB size to tar.gz

Say /var/log/tes directory has a file with 10 MB I want it to compress to tar.gz

View 2 Replies View Related

General :: Download A Single File In 2 Parts To Different Locations Using Wget?

Jan 18, 2011

I need to use wget (or curl or aget etc) to download a file to two different download destinations by downloading it in two halves:

First: 0 to 490000 bytes of file
Second: 490001 to 1000000 bytes of file.

I will be downloading this to separate download destinations and will merge them back to speed up the download. The file is really large and my ISP is really slow, so I need to get help from friends to download this in parts (actually in multiple parts)

The question below is similar but not the same as my need: How to download parts of same file from different sources with curl/wget?

aget

aget seems to download in parts but I have no way of controlling precisely which part (either in percentage or in bytes) that I wish to download.

Extra Info

Just to be clear I do not wish to download from multiple locations, I want to download to multiple locations. I also do not want to download multiple files (it is just a single file). I want to download parts of the same file, and I want to specify the parts that I need to download.

View 1 Replies View Related

General :: Splitting Text File Into Several Parts By Line Count

Sep 21, 2009

I have a utility that works with files. The utility is crashing at after about 120 files. The input to the utility is a file containing a filelist. I want to cut the file with the file names in it to seperate files containing about one hundred or so. My thought was to determine the number of lines/100 and then use head and delete to create temporary files to run the utility multiple times to prevent the crash. When I tried to create a variable using the wc -l command the output gives me the number of total lines but it also includes the filename of the input file. (873 Filename.txt) I can not figure out how to remove the Filename.txt from the variable.

View 2 Replies View Related

General :: Split A Video File Into Smaller Chunks?

Sep 1, 2011

Tool like Media File Segmenter for Linux

I know that one can use ffmpeg to extract a smallfile.avi from a largfile.avi. But What I am looking for is an tool/command to split a large file into several files of a given size.

View 1 Replies View Related

General :: View A Particular Ten Lines In A Large File Where Can't Open The File In Vi

May 12, 2010

I am using RHEL 5.I have a very large test file which cannot be opened in vi.The content of the file has some 8000 lines.I need to view ten lines between 5680 to 5690.How can i view these particular lines in a large file.what is command and option i need to use.

View 1 Replies View Related

General :: Best Way To Copy A Large File Over NFS?

Aug 24, 2011

I want to transfer a huge file (60GB) over the NFS network on linux. Is cp the best option?

View 1 Replies View Related

General :: Can't Copy Large File?

Mar 26, 2010

I'm trying to copy a 6Gb file across from my laptop to an external usb drive but it quits at about 4.2Gb every time with a "file size limit exceeded" error. i have checked the output of ulimit -a and there is no limit there on the file size. I'm using the Slax live Cd for this as it always gets the job done

View 8 Replies View Related

Ubuntu :: Check The File's Compress Method?

Feb 1, 2011

I have a compressed text file. The method of compress is unknown.I can see the file contents by using Midnight Commander without a problem but I would like to view the file just with cat. So I am trying to uncompress the file with unzip or gunzip but it does not work.How to check the method the file is compressed with? Is any way to find it with Midnight Commander?

View 4 Replies View Related

Ubuntu :: How To Compress Huge File With JPEGs

Feb 23, 2011

I want a huge file of .jpeg files. How can I can compress these files to save my space? Is there any s/w for this??

View 2 Replies View Related

General :: How To Send A Large File Securely

Aug 28, 2011

I need to send large files from a Linux machine to another using cryptography. The sender machine knows the recipient IP but not vice-versa. I don't need strong cryptography and prefer higher-speed less-secure solutions.

There are no problems with presharing crypto keys but I'd prefer not dealing with SSH users creation.

I think to HTTP PUT over TLS, but I never had experience with it and I prefer to hear which are the possible solutions. I know that it can listen as a daemon but I don't know anything about cryptography. So pipeing with OpenSSL may be a solution.

View 2 Replies View Related

Fedora :: Wavmerger \ Merge Them And Compress Them To Create A Mp3 Or Ogg File?

Jun 18, 2011

have 6680 wav files with about 500kb size in a folder and i want to merge all of them.the size of the files altogether is 1.5GB. how i can merge them and compress them to create a mp3 or ogg file?

View 4 Replies View Related

Ubuntu :: 10.04 LTS Lucid - Root Is Getting Smaller And Smaller

Jan 22, 2011

I am using 10.04 LTS Lucid, and I notice the free space of root is getting smaller and smaller.

Five months ago, there was about 3.9GB free space of root, but now it is only 1.6GB. I always run sudo apt-get autoremove and sudo apt-get autoclean every time the update is finished, and also use Bleachbit to clean the system, but both are useless.

I never faced such problem with older versions of Ubuntu, is there any measure to fix it?

View 6 Replies View Related

General :: Using Sed To Replace A *large* Number Of Variables In A File?

Jul 28, 2011

I have a large number of log files, on a linux box, I need to cleanse sensitive data from before sending to a third party. I have used the below script on previous occasions to perform this task, and it has worked brilliantly (script was built with some help from here :-)

#!/bin/bash
help_text () {
cat <<EOF
Usage: $0 [log_directory] [client_name(s)]
EOF

[Code]...

However, now one of our departments has sent me a CLIENT_FILE.txt with 425000+ variables! I think I may have hit some internal limit. I have tried splitting the client file into 4 with around 100000 variables in each, this still doesn't work. I'm loathe to keep splitting though as I have 20 directories with up to 190 files in each directory to run through. The more client files I make, the more passes I have to do.

View 2 Replies View Related

General :: Monitoring Copy Progress Of A Large File?

Sep 15, 2010

Is there a clever way to monitor the progress (as percentage or hash) of copying a large file (using pv could be an option)?Like monitoring the progress of a copy command such as this:Code:cp linux.iso /tmp/

View 2 Replies View Related

General :: Backup Large File To Multiple DVDs

Nov 2, 2009

I work for a school consulting company.We helped a school deploy about 1500 computers.The computers have windows XP but we have been using G4L for the restore partition on the drives.So far the software works great. We did however run into a problem in that many of the computers we deployed are missing the restore partition. The reason they are missing is long and convoluted and not really that important. What I have been charged to do is try and fix the restore partition problem. One solution that I had, which im not even sure if it will work, was to backup the recovery file, that g4l created, to DVD and write a basic script to recreate the partition and then copy the file over. This process would need to be as automated as possible since this disc will be inserted by the end user(the students). The backup file that g4l created is 5.9GB so it wont fit on just one disc and Dual layer discs are too expensive to use for this project, so the file will either need to be compressed again (not sure if that's a good idea or not) or split across two DVD's.

I have searched the forums here and I was not able to find anything to fix this problem. I was able to find some info on splitting files across two discs but im not sure how to use that to fix my problem.

View 5 Replies View Related

General :: Cannot Find Large Untared MySQL File

Aug 6, 2009

After untaring a mysql file (very large) I'm trying to find where the file listed below has gone. I did a search on the file name:
fine / -name 'mysql-qui-tools-5.0' -print
But can't find the file.
-rwxr-xr-x root/root 9651820 2007-05-02 11:46:01 mysql-gui-tools-5.0/mysql-query-browser-bin

View 6 Replies View Related

Debian :: File Upload Using Firefox - Smaller Image File Is Previewed While Bigger Image File Is Not

Dec 16, 2015

I am trying to upload some pics on my Facebook account using Firefox. When I click on Facebook's file upload icon, Firefox bring up a 'File Upload' window. I noticed that smaller image file is previewed on the lower right hand corner, while bigger image file is not. Is there anyway I can change this behavior or maybe change what Firefox is using to browse my files?

View 2 Replies View Related

Ubuntu :: How To Compress / Reduce Resolution Of Tiff Image File

Jan 4, 2010

how do I compress, or reduce the resolution, of a tiff image file.I must use tiff I can not convert to JPEG etc.

View 3 Replies View Related

Software :: File Extractor (rar, 7z, Ect) For Part Files Compress By Winrar?

Apr 12, 2010

I download a lot of files that have been compressed into separate pieces by winrar, so that a 4gb file will be shown as dozen or so 93mb peices. In windows, using winrar, extracting them was rather easy but in Kubuntu whatever default program is being used throws a fit. In a perfect world I would want to get a open source alternative from synaptic, alternatively anything I can download in RPM I can eventually figure out. I want to avoid trying to get winrar working under wine.

View 13 Replies View Related

General :: Move Large Amounts Of Music Within A File Structure?

Dec 20, 2009

i have a car stereo that reads a USB drive with all my music on it, however to sort through the music it uses a method of finding folders containing music, then displaying them all in a list. i find this interface annoying because in order to sort the music by artist i have to go and manually move it out of the album folders by hand, this takes a long time for 11+ GB of music so i was trying to use the linux CLI to quicken the process. use a command like this

Code:

mv /media/usb/music/*/*/* /media/usb/music/*/

but for some reason this moves all my music into the last folder alphabetically in my drive, the music is all pre-arranged like this /media/usb/music/artist/album/song

View 5 Replies View Related

General :: Split A Large File To Download On A Windows Machine?

Apr 16, 2011

I am removing some old graphics from my server and one of the gallery programs have created two enormous directories that I cannot even open with FTP.

I tried to tar each directory and the first came out to about 37gb and the second keeps failing (its bigger one would assume).

How can I archive and split these into smaller files?

View 13 Replies View Related

General :: Accidentally Type Concatenate Large File On Remote System

May 6, 2010

Every once in a while on a computer I'm ssh'd into, I will accidentally type "cat largefile.txt" and my screen will start rushing with text for the next 10 minutes. I'm always working in a screen session, so my current solution is to just log out and then log back in, and since it can go 100X faster when I'm logged out, it'll finish in the short time it takes me to type my password in again. Is there a better way? Either involving the fact I'm in a screen session? Or a way to do this within SSH? What doesn't work: detaching from the screen session (doesn't respond until file is done outputting) trying command to move to a different window in the screen session (also doesn't respond) typing ctrl+C to kill cat command (also doesn't respond, probably because the command is done and the buffers just have to catch up).

View 3 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved