Software :: Bash Script To Extract Multiple Files With Password?
Jan 10, 2011
The situation is the following:I download everyday several files from a website which are always password-protected with one out of 4 or 5 different passwords.I wanted to write a bash script to extract them automatically by trying with the different passwords.The files are mostly 7z.Can anbody lend me a hand?The starting point could be sth. like:
I'm trying to write a script that will prompt the user for a username/password, then create that user/password in the right groups on all my machines. I know this is kind of a long way around to avoid a NIS server, but I like making my life more difficult.
This is what I have so far:
the script has 2 problems: The "if" functions return an error and do not compare the strings successfully. whatever password is entered does get applied properly and the user is unable to login
I have wrote a 1 line command that parses a file, locates the IP Address in the file and then trims the output the way I want it, and then sorts numerically and by uniqueness and then >> appends to output.txt
I can get all the IP's into 1 file "output.txt", but what I am really looking for is some type of way to create a text file, for each IP it finds labeled xxx.xxx.xxx.xxx.txt and also put that ip address into that file..
I have this cool bash script that I worked hard on. But it broke down when it can across files that had non-English characters. Another small problem was getting it to descend into a directory. If it renamed a directory it would not descend into that dir to rename the other files. I would have to run the script twice on the same directory.
Here is the script: Code: find -type d -o -regextype egrep -iregex '(.*.ogg|.*.mp3|.*.wav)' | while read s do rename -v 'y/A-Z/a-z/' "$s" done find -type d -o -regextype egrep -iregex '(.*.ogg|.*.mp3|.*.wav)' | while read n do rename -v 's/ /_/g' "$n" done A French name like this:
Code: Chateau De Sable (imagine accents above the letter a) became this:
Code: ch303242tea_de_sable This is not what I wanted.
Why would the script not descend into a directory after it was renamed?
I want to remove duplicate or multiple similar lines from multiple files. I.e. if I have four files file1.txt file2.txt file3.txt and file4.txt and would like to find and remove similar lines from all these files keeping only one line from these similar lines. I only that uniq can be used to remove similar lines from a sorted file.
I have written quite a few separate bash & scripts and php scripts that up to now I have run from cron jobs. However I have to estimate how long each takes to run, before running the next and so it probably takes much longer than necessary to run them all. They have to run in order.
Now there are so many I am thinking it would be better to have a master bash script that would run one after the other, but I am not sure how to get the master script to wait before starting to run the next script. Is this possible and is there a command that will make the script wait between bash and php scripts , for them to finish, before running the next?
I'm not sure if this is possible or even where to start. I assume that this can be done with an sh script using tar or similar.I have several very large zip files that contain images for all of the products in my online store. Each image is named after its 13 digit SKU (for example, 9987788000012.jpg). In order to import products into my store, all images are placed into a media directory. Unfortunately, there are over 100,000 images.
So I would like to break the images into sub-folders based on file name. For example, when I extract store_images.zip (or tar or whatever), my extract script would create directories (if they don't already exist) based on the first three digits of each image name, placing each image into the appropriate bottom level directory. For example, "9987788000012.jpg" would be placed in the following directory "media/9/9/8", with media as the root and "8" as the directory that holds any images that start with "998". Perhaps two sub-folders would be less cumbersome.Assuming this requires a script, particularly since it involves scanning image names, creating folders, and saving images to specific directories, which language would serve my needs best? PHP? Has anyone had to do something similar?
I have two video files (Xvid) and would like to combine the video from one of these with the audio track of the other, in order to create a new video file.
This is somewhat complicated by the fact that I would like the resulting audio to be a mixture of the two original audio tracks, for instance, during some time segments, I would like to switch from one to the other, but the video should always be the same.
Another issue that complicates the things is that the two audio tracks have different bit rates, and when I briefly managed to merge the two, one of the audio tracks was playing much faster than the other. To clarify, the audio tracks should not overlap but just be played at the different time during the video playback.
I am trying to do this by using Audacity. The problem is that I am fairly new to Audacity and I have not been able to find any info in their user guides regarding this specific issue.
Archive: /home/kishore/Desktop/wrar391.exe [/home/kishore/Desktop/wrar391.exe] End-of-central-directory signature not found. Either this file is not a zipfile, or it constitutes one disk of a multi-part archive. In the latter case the central directory and zipfile comment will be found on the last disk(s) of this archive. zipinfo: cannot find zipfile directory in one of /home/kishore/Desktop/wrar391.exe or /home/kishore/Desktop/wrar391.exe.zip, and cannot find /home/kishore/Desktop/wrar391.exe.ZIP, period.
i am not able to extract the files in ubuntu can you tell me when i click archieve manager i am getting this error and which software i have to download for the zip files to be extract in ubuntu
I haven't been to this site in quite a while, since it changed from LNO in fact. Good to see this place is still around, albeit under a newer name.
I'll get on with the problem. I've got a Netgear SC101T that I was using to store my files on. Some of you may know it uses the DataPlow SAN file system. It worked fine until I installed a firmware update which, for some reason, broke the mirror array. I've hated this POS ever since and want to pull the data from the drives and toss the box. The problem is, linux doesn't have support for this particular file system scheme.
What I'm wondering is, how does 'dd' work, in regards to keeping the file system. Does it simply copy files and disregards the structure, or does it make an exact copy, DataPlow FS and all? Anyone else ran into this conundrum?
I have a file : cpq_cciss-2.6.20-34.rhel4.i686.dd which is designed to build a floppy disk; these floppy is used to hold disk driver which is not on RedHat CD-Rom. But this .dd is not complete: some files, like /drivers/pci.ids are missing.My idea is to extract all files from .dd file, put missing files and then re-create a new dd file. But, how can extract all files from initial .dd file, and then recreate a new one?
I'm trying to extract .part files. These are mac extensions similar to .rar or .zip. I need to extract a whole bunch of them simultaneously. Theyre probably registered in sequence just like .rar files. What program can I yum install or where should I look to extract these to what will probably be an ISO.
I have many files in a folder from which I need to extract some contents, these are basically text files wich have individual lines with (i.e)
name: john address: whatever phone: 123456
1. Sometimes a line might be missing.
name: johnn phone: 123456
2. Lines are not in the same line-numbers across the files I did try some things with awk based on google searches but I couldn't extract the data of each file into a single line (this is the ultimate goal):
I don't have knowledge other than having put some bash scripts together for backup jobs, so I am open to install anything that could to pull this off.