General :: Write A Script To Compress Any File With 10 MB Size To Tar.gz?
Jun 7, 2010write a script to compress any file with 10 MB size to tar.gz
Say /var/log/tes directory has a file with 10 MB I want it to compress to tar.gz
write a script to compress any file with 10 MB size to tar.gz
Say /var/log/tes directory has a file with 10 MB I want it to compress to tar.gz
i am using classpath-0.98,jamvm-1.5.4 and arm9 cortex processor. so my question is after install classpath and jamvm on arm9 , i am getting around 30MB FAT file. so tell me some tips how to reduce the size of FAT file as much as possible.
View 1 Replies View RelatedI am trying to compress a folder by right clicking and selecting Compress...I get the following error:
An error occurred while adding files to the archive.No such file or directory
I want the folder and its content to be compressed to a .ZIP file which is natively accessible by Windows.
I'm looking for a way to compress a large file (~10GB) into several files that wont exceed 150MB each.
Any thoughts?
I've a file with a size of 6GB. I would like to compress this file and split them into smaller files. I was also thinking in use bzip2 to compress it, because if offers a good compression rate. How can I split this file into small ones to compress it?
View 5 Replies View Relatedi need to write a short script that will compress a specific folder that`s on the Desktop (and all it`s content) and also will encrypt it with a password that is inside the script --->meaning it wont ask for a password+verification when compressing+encrypting
View 1 Replies View RelatedI was attempting to reformat a 16GB MicroSD card in my camera when the battery died mid-way. After that, any time I try to read the card in my camera, it gives me a "Card Error" and does not allow me to reformat it in my camera.
So, I thought I would plug the camera in to the laptop with it set to host the card as media when plugged in as USB, in an attempt to fix the formatting issue.
However, when I plug it in to my linux machine, it does not register as a device (e.g., /dev/sda) due to some errors, therefore I cannot reformat it. Essentially, I think I need to fix the partition table but I'm not sure how to when it doesn't register as a device. code...
I have a compressed text file. The method of compress is unknown.I can see the file contents by using Midnight Commander without a problem but I would like to view the file just with cat. So I am trying to uncompress the file with unzip or gunzip but it does not work.How to check the method the file is compressed with? Is any way to find it with Midnight Commander?
View 4 Replies View RelatedI want a huge file of .jpeg files. How can I can compress these files to save my space? Is there any s/w for this??
View 2 Replies View Relatedhave 6680 wav files with about 500kb size in a folder and i want to merge all of them.the size of the files altogether is 1.5GB. how i can merge them and compress them to create a mp3 or ogg file?
View 4 Replies View Relatedhow do I compress, or reduce the resolution, of a tiff image file.I must use tiff I can not convert to JPEG etc.
View 3 Replies View RelatedI download a lot of files that have been compressed into separate pieces by winrar, so that a 4gb file will be shown as dozen or so 93mb peices. In windows, using winrar, extracting them was rather easy but in Kubuntu whatever default program is being used throws a fit. In a perfect world I would want to get a open source alternative from synaptic, alternatively anything I can download in RPM I can eventually figure out. I want to avoid trying to get winrar working under wine.
View 13 Replies View RelatedAre there software that can split big file size into small file size in Linux?
View 1 Replies View RelatedI was just testing specifying limit on file size to a user and have added the following to /etc/security/limits.conf bob soft fsize 100 This basically should have said not to allow bob to create anyfile greater than 100Kb in size.
But the interesting thing is, if bob already has any file which is greater than 100Kb in size, it even doesn't allow to log him into the system both from console and SSH. Also nothing is logged in logs.. How do I configure it so that, bob can login to the system even though he has any file greater than 100Kb (but doesn't allow him to create file which are greater than 100Kb) ??
When I ls -l /etc/passwd, -rw-r--r-- 1 root root /etc/passwd When I login as myself, and rm /etc/passwd, it asks: rm: remove write-protected file '/etc/passwd'? If I say yes, will it actually delete the passwd file?
View 1 Replies View Relatedis lvresize with --resizefs options re-size the Logical Volume and then re-size the file system? i mean we don't need to use resize2fs?I looked at man pages but it doesn't explain this option.
View 3 Replies View RelatedHow can we find the maximum size of the inode table and what decides it, and how the maximum size of volume of file system is decided ?
View 4 Replies View RelatedIve installed Gaussian '03 on fedora Core 10, but I'm unable to run it. It aborts and i get the following error
Code:
Erroneous write during file extend. write -1 instead of 4096
Probably out of disk space.
Write error in NtrExt1
I have a large list of folders that I want to compress using tar.the script I want to right is to travers all the folders and for each I would tar the contents to a tar file with the same name of that folder. and the it should be stored inside the same forlder.
View 6 Replies View RelatedI'm trying to write a script that will calculate a directory size and if the size is greater than 4GB, it will send out an email. But I am getting the "integer operation expected" error when it tries to calculate the current disk size. Here's my script.
Code:
TO="user@email.com"
SUBJ="Ready for DVD burning on `hostname`"
MAIL="/bin/mail"
[code]....
I want to package my application for linux, I use zip/tar.gz/7z/rar to do the compress in Windows.Because I do the compress in Windows, so when I extract the package in Linux, the file permission is reset.The problem is Only the tar.gz can let me extract runable files, other format (zip/7z/rar) not (the file permission is reset to 644).So my question is how to compress my files using zip/7z/rar while reset the permission to 755.
View 2 Replies View RelatedI'm writing a script/plugin for Nagios for testing a WebLogic server.. I redirect some output to a file, and then i read that file to get some data, but i can't seem to write to that file with my script :s... this is the most important code
[Code]...
* EDIT * When i execute this script through a local terminal (PuTTy), it works but when i execute it from Nagios, it doesn't work..
I have a large number of folders that each contain quite a few files of varying sizes (from a few bytes to 400kb or so), mostly smaller ones. I need to get the actual (not the disk usage) size of these folders. Is there any way to do this with a command like 'du'?
View 4 Replies View RelatedThere seems to be a limit of 2 gb on files. Im trying to add my music files and have about 7 gb. Is there a way to make a file that will allow the extra info.
View 7 Replies View RelatedFirstly, I did perform a search on this problem in these forums, but didn't quite get what I was looking for. So I hope I don't yelled at for making a duplicate post. So I used rsync to backup my webroot to another nix machine. du -hs gave me 1.3 G on the source machine and 1.1 G on the backup machine. I tried to compare the individual files and noticed a trend. The files on backup machine were always smaller than the files on source machine. The source uses SATA drive, destination uses IDE. So this time I rsynced locally to another folder on the source machine. Same size anomaly. So i did a simple cp file ~/file and same size anomaly. So it's not a rsync issue.
I took a file and ran md5sum on both, the source file and destination file. To my surprise, even though the file size was different, they had the same md5sum. Now, let it be known that the source machine is a production server and the dir i rsynced was being used, serving pages to the web. I googled about this and came up with stuff like open descriptors and holes. I don't understand this stuff and was wondering if this was really the case. What are those if it is the case? And my backup copy is 100% identical right? There are thousands of files and I ran md5sum only on couple. Can I take comfort that when time comes, I can restore using my backup without any problems?
I found one weird behavior.The file i copied from a server is almost doubled the size from the server.How come?
example:
720K ./xxxx vs 360K ./xxxx
(my pc) vs (server)
I have checked both filesystem are ext3
How to reduce the size of a file. Is truncate related to this?
View 12 Replies View Relatedi have this directory with multiple images 'pics' and the size is 20mb, and i want to make a .zip or .rar package of this directory but with an increased size so the .zip/.rar file will be 100mb, and then when you extract it the file size is the original 20mb
View 10 Replies View RelatedI want to keep a trace of the URL I visit, so I use a command line like this:
tcpdump -ien1 -v -X 'tcp port 80' | sed -nl
's/^.0x[0-9a-f]{4}:.{43}(.)$/1/p' |perl break.pl |perl -pe
's/(GET|POST).(.*?).HTTP/1....Host:.([a-zA-Z._0-9-]*)../"
BEGURL
[Code]....
I also tried redirecting stdout and stderr to /tmp/out, it's still empty. The file has write access. I have no idea what it can be. Is there anything else than stdout and stderr?
I want to write a shell script file for the below subject
subject / situation : i have many users say user1, user2, user3, user4 and so on... within my /home dir
Within a user dir say.. /home/user1 i have many unwanted files. these unwanted files start with the name core for eg. core2324, core9789, core 9079 etc.. i need to delete them.
I want to write an automated script for this, which can do the same. How to write a script which can delte these unwanted core files which exist in all user dirs.