General :: Gzip Multiple Files Into One Gz File?
Mar 29, 2011I have 100 files: cvd1.txt cvd2.txt ... cvd100.txt
How to gzip 100 files into one .gz file so that after I gunzip it, I should have cvd1.txt, cvd2.txt ... cvd100.txt separately?
I have 100 files: cvd1.txt cvd2.txt ... cvd100.txt
How to gzip 100 files into one .gz file so that after I gunzip it, I should have cvd1.txt, cvd2.txt ... cvd100.txt separately?
I have a backup script basically is this
Code:
BACKUP_DIRS="/etc /boot /root /home"
BACKUP_FOLDER="/tmp/system_backup/
for DIR in ${BACKUP_DIRS}
do
[code]....
All the folders get dumped into seperate gzip files. Now I want all the gzip files in the backup folder into one final gzip or bzip2 file. My goal for this is to get one file instead of multiple so I can scp or ftp the one file to another file share. Which would be easier to send one file than a bunch of files.
I have installed debian on my dell latitude d630 laptop and am currently making a derivative for others to use. however I can't get the wireless to work. I downloaded it from the broadcom site but it comes as a gzip file instead of a deb file. if I extract the gzip file I don't get any installable (deb) files. I need the wireless to work because now a bridged network from my windows laptop isn't working.
View 3 Replies View RelatedI have a existing zipped file , I want to use gzip command to append some files to it , I tried man gzip but can't find the key word "append" , can advise how can I do it ?
View 1 Replies View RelatedI am going crazy with a gzip file. I can decompress the file in Windows using WinRAR but it is impossible on any UNIX operating system. the file seems to be ok. If I do file the_name_of_the_file.gz
I get: the_name_of_the_file.gz: gzip compressed data, from Unix, last modified: Sun Jan 30 14:10:21 2011
But if I do gunzip -f the_name_of_the_file.gz I alsways get: gzip: the_name_of_the_file.gz: unexpected end of file The same problem happens when I try to extract the file using the GUI tool in Ubuntu or MacOSX,
I have a script which periodically backs up a directory using the command "tar -czvf [name] [directory]" but my problem is that the script has recently been putting a lot of stress on the server (Minecraft SMP) and tends to lag players as it backs up, which recently has been taking nearly 5 minutes.So I need to know if there's a way to control the GZip compression rate at the same time that it archives and backs up the files?I understand that I can first tar the files and then GZip them separately with a different compression rate afterwards, but this would not work because it names the files with the current server time, which sometimes changes in between commands.
View 1 Replies View RelatedI have a file with 5 columns. Column 4 contains numbers.Is it possible to split the file into multiple files using a condition for the contents of column 4 i.e if column 4 contains a value between 0-10 then print the lines to a new file called less_than_10.txt
View 1 Replies View RelatedI am trying to upload an IOS in the cisco NAC Appliance. The IOS version has to be updated as 4.8. I am getting the below error when i tried. File is not in gzip format Child return status 1 Error exit delayed from previous errors. I am using the below command to unzip the IOS file. tar xzvf ccca_upgrade-4.8.0-from-4.6.x.tar.gz.
View 3 Replies View Relatedhave a gzip file ABC_000023232.gzipBCD_023232032.gzipI want to split these files into smaller files but keep the extension same because I am using this as a variable in a script
Code:
for i in `ls *.gzip`
split -b 500K $i $i
[code]...
I wanted to be as clear as possible. I have to add hundreds of new servers to hundreds of configuration files in Nagios. Here is a sample structure:
servers.txt has:
SERVER3
SERVER4
[code]....
I have a file which contains the data i retrieved through prstat and an array that contains all the unique process ID's of that particular file. i want to compare each and every line in the file with each and every element of the array so that i can create multiple files for the multiple value in the array.
View 6 Replies View RelatedI'm working on some scheduled task script files to keep nightly backups of some of our database information in place, and it's a bit annoying when they blow up. I know how to redirect stdout and stderr to a flat file I can view when I come in, and I know that 2>&1 maps them both to the same file (whatever was named in 1). However, I'm running into some cron-time situations where it's easier to have the two streams together, and other cron-time situations where it's easier to have them separated. I can't really tell which is going to happen; is there some way I could create both kinds of output file for my scripts, so that I've got a std_err only file and an interleaved std_out/std_err file?
Note: I've looked at the 'tee' command, but I don't think it will work for what I'm after. 'tee' appears to only work with stdout; I'm trying to work with stderr.
I am using Xfce as the desktop enviroment and Mozilla Firefox as the webrowser. Within the webrowser window, I do File>Save Page As. I save it, and the result is almost always foo.html and directory foo_files. But I think under KDE I could choose the format, one of them being something like "Single page" (only one file; the colecction of .png, etc is embedded into that file). And this is the format I want Xfce (or Firefox) to use when downloading to hdd.
View 1 Replies View RelatedBeing new to this area .I have been assigned a task which i am unable to do . Can any one please help me .
I have requirement where i have input file XYZ_111_999_YYYYMMDD_1.TXT and with header and series of Numbers and Footer.
I want to create a mutiple output files with each file having a seperate code which is stored in text file and create XYZ_222_999_YYYYMMDD_1.TXT . and add date in the contents next to series of numbers .Like this
Before change the file looks like this
file name XYZ_111_999_YYYYMMDD_1.TXT
001,19SEP2009-14:05:05,000000003
1234
4567
6785
END_OF_DATA
[Code]....
A function by name abc is called in many files. I want to copy all the lines with the function call to an output file.A simple grep on function name doesn't help me as the function call is spanning across multiple lines as follows:
abc(parameter1,
parameter2,
parameter3);
So I want to copy all the three lines (till semicolon) to the output file.The problem is because there are more than 200 calls for the same function and I cannot do it manually
I have limited access to several servers (key based auth) but cron facility is not available for me. Those servers getting filled up by large apache logs and I have to login to each node manually and clean them each and every day.
I tried to write a script to run from login box but when i try that it looks like it is looking for logs in the local server (login box).
So current situation is:
How can i modify this so that the script in server1 will look for files in that server and zip them?
Google showed another command called rsh but in my env it is also not avil.
I am to rename all the files within a directory (which contains multiple subdirectories) recursively without invalid characters.
I tried the coding posted above.
find . -type f -printf '%p
' | while read file; do
oldfile=$(basename "$file")
newfile=$(echo "$oldfile" | sed 's/[^A-Za-z0-9_.]/_/g')
if [ ! "$newfile" == "$oldfile" ]; then
echo mv "$file" "${file%$oldfile}$newfile"
code....
but I get an error on both of them stating "find: bad option -printf find: [-H | -L] path-list predicate-list"
file roller no gzip bz2 password supported?
View 2 Replies View RelatedI have downloaded apache(file name: httpd-2.2.17.tar.gz.bz2) throughw wget at /usr/local/src now i m not able to do next step with gzip command I am using following command and getting error
]# gzip -d httpd-2.2.17.tar.gz.bz2
gzip: httpd-2.2.17.tar.gz.bz2.gz: No such file or directory
Using lucid LTS, I have no problem installing dropbox using either the website download .deb, or adding it to my repositories.
The problem comes when I try and run the software. At this point it unsuccessfully tries to download the required proprietary daemon.
In the terminal, running:
'sudo dropbox start -i'
Gives:
Starting Dropbox...Traceback (most recent call last):
File "/usr/bin/dropbox", line 242, in handle_data_waiting
self.unpack_dropbox()
File "/usr/bin/dropbox", line 252, in unpack_dropbox
[Code]....
So im trying to un rar my rar. files. I use 7-zip, and winrar, Every time i try to unrar them it says broken file, cannot open file. This happens to nearly most of my rar files in the last few weeks, idk what else to do,
View 3 Replies View RelatedI am looking for a program that will take multiple avi files and join them into one file
View 8 Replies View RelatedI have a log file on ubuntu 10.04 that has 500 lines of log data in it. What command could I use in a terminal to split the single 500-line file into generate ten files each with 50-lines of log files each?
View 2 Replies View RelatedI have made the simple mistake of using dolphin to rename several picture files of types jpg and png. I have no problem viewing them and Ubuntu still knows which is which (if i view their properties it will state File Type: JPEG or PNG) but unfortunately Ubuntu Tweak does not allow you to chose them without the extension (I am trying to change my login background using this) So what i need is something that can scan several files, determine the file type, and add an appropriate extension to it. I have found several to do so for music files but none for images. I am using Ubuntu 11.04 at the moment.
View 5 Replies View RelatedI've been trying to sort this out for several hours and I?m totally lost? I?ve been searching around, but haven?t found the solution to my problem. I have a directory with 100 files. I need to copy 10 lines of each files (let?s say from line 45 to 55) into one unique file. So I guess I could use sed ?w, but I didn?t manage to write the right script. I also tried using a loop to create 100 different files, each one with the 10 lines) to concatenate them later on. But I only got 1 file, not 100.
View 12 Replies View RelatedHow can you create a script to move or copy files from a main directory into multiple directories below the main directory.
View 7 Replies View RelatedHere is the deal: my school purchased a test prep which they distributed to us, the only problem is that it is for windows. When I try to run the exe in wine the program windows pops up but the screen is blank (i can see the menu bar and and the title bar). It appears that most of the information is located in a folder as html files (separated into chapters for each section) with links to all the figures/tables/questions
(also html files). This morning I have been trying to find a way to compile the html files into a format that I can read without having to hunt for each individual html file. I tried navigating with firefox but had to change each individual link before I could do that (very time consuming), I tried using doxygen (the GUI front end) to convert it to Latex/rtf/html/xml (all the formats) but did not have any success (It could be that I am using the app wrong). I tried opening some java applications found within the application data folder in hopes that it would bring up the navigator without success. And even tried installing java in wine (unsuccessfully) to see if that was the missing key to run the program in wine.
My question is, do I just have to fork out the $500+ that it would cost me to purchase the books instead of using the school provided program (free to us)? or is there a relatively simple way to compile these files into a format that could be readable.
tl;dr - have multiple html files for a windows native program that won't run on wine that I would like to convert to a more easily readable format (i.e. chm, pdf, etc), is it possible?the program was legally acquired through my school, I did not download or pirate it. I am not looking to distribute it, just looking to use it for my own studying purposes without spending $500+ or buying a windows license/computer.
Given a gzip compressed file, how do I know what compression level (1-9) was used for it?
View 2 Replies View RelatedI want to be able to write a shell script for downloading files (only *.tar extension) from multiple folders (the sub folder's names may vary) in a FTP site and be able to untar them and then gzip them and then move them to the real folder.
View 4 Replies View RelatedDoes gzip have the capability to decode gzipped traffic? I have been beating my head against the wall with this issue. What I'm trying to do is capture traffic between a web server and clients, and I've got it set up where it's redirected to a file for ease of grepping, however it's seemingly incapable of decoding gzipped encoding. I know I can do this with tshark, I'm curious as to whether tcpdump has this capability (i.e. only using tcpdump, and not some additional tool like tcpshow or what-not).
I can't find much on this issue in the man page for tcpdump, but it is fairly lengthy, so it's possible that I missed something, but I don't see that as especially likely.