Red Hat / Fedora :: Extract Files From A .dd File?
Jul 28, 2010
I have a file : cpq_cciss-2.6.20-34.rhel4.i686.dd which is designed to build a floppy disk; these floppy is used to hold disk driver which is not on RedHat CD-Rom. But this .dd is not complete: some files, like /drivers/pci.ids are missing.My idea is to extract all files from .dd file, put missing files and then re-create a new dd file. But, how can extract all files from initial .dd file, and then recreate a new one?
I have installed debian on my dell latitude d630 laptop and am currently making a derivative for others to use. however I can't get the wireless to work. I downloaded it from the broadcom site but it comes as a gzip file instead of a deb file. if I extract the gzip file I don't get any installable (deb) files. I need the wireless to work because now a bridged network from my windows laptop isn't working.
I'm trying to extract .part files. These are mac extensions similar to .rar or .zip. I need to extract a whole bunch of them simultaneously. Theyre probably registered in sequence just like .rar files. What program can I yum install or where should I look to extract these to what will probably be an ISO.
I have a .tar archive of music that I would like to extract specific files / folders from. I've been looking at man tar but I don't seem to be able to find any information related to my quest, although someone will likely point out that it's right there in front of my nose
Anyhow, the file is called music_archive.tar and I would like to extract the folder /music/P/Pink Floyd/The Wall/. Anyone know how to do this
I'm trying to figure this error message out. This little script is supposed to tweet my laptop's IP address, as a cron job, I'm hopeful that it would do so even if it's stolen. This is a variant of one that works, but this doesn't, and I can't see a difference in the curl line of either one.
Code: #!/bin/bash user="xxxxxx@xxxxxxxxx" pass="xxxxxxxxxxx" wget [URL] TWEET=`sed -n 1p index.html` curl --basic --user "$user:$pass" --data-ascii "status=$TWEET" "[URL]" rm -f index.html exit This is the error message.
Code: curl: (6) Could not resolve host: status=66.183.103.67; Cannot allocate memory {"request":"/statuses/update.json","error":"Client must provide a 'status' parameter with a value."} Why does curl think the status is the URL?
I have a multi-sector nrg file that I would like extracted. I can't seem to find a way to extract the contents of it! Please tell me if there is a tool I can use to do this!
Why do I always need to write "su" then my password to extract or copy any file in fedora 11. How to configure so that I always be in my root directory.
When ever I extract a file from a .tar or whatever, it isn't detected. I notice this mainly when i'm using xampp. I copy zip up all my files on one computer, load ubuntu on another, extract the files to the web folder (htdocs) and then I get nothing. However, when I manually create the files directly on my computer as opposed to extracting them, they appear.
Is there something I need to do in order to have these files appear? Is there some sort of file system refresh? Or am I being a complete idiot?
Archive: /home/kishore/Desktop/wrar391.exe [/home/kishore/Desktop/wrar391.exe] End-of-central-directory signature not found. Either this file is not a zipfile, or it constitutes one disk of a multi-part archive. In the latter case the central directory and zipfile comment will be found on the last disk(s) of this archive. zipinfo: cannot find zipfile directory in one of /home/kishore/Desktop/wrar391.exe or /home/kishore/Desktop/wrar391.exe.zip, and cannot find /home/kishore/Desktop/wrar391.exe.ZIP, period.
i am not able to extract the files in ubuntu can you tell me when i click archieve manager i am getting this error and which software i have to download for the zip files to be extract in ubuntu
I am running Ubuntu 11.04 and just tried to install VLC. I download the .zip file and once it loaded, I tried to extract it. The following error came up:
tar (child): /tmp/vlc-1.1.4.tar.bz2: Cannot open: No such file or directory tar (child): Error is not recoverable: exiting now tar: Child returned status 2 tar: Error is not recoverable: exiting now
What is the correct method to unzip and install this (or any other .zip file for that matter?)
I haven't been to this site in quite a while, since it changed from LNO in fact. Good to see this place is still around, albeit under a newer name.
I'll get on with the problem. I've got a Netgear SC101T that I was using to store my files on. Some of you may know it uses the DataPlow SAN file system. It worked fine until I installed a firmware update which, for some reason, broke the mirror array. I've hated this POS ever since and want to pull the data from the drives and toss the box. The problem is, linux doesn't have support for this particular file system scheme.
What I'm wondering is, how does 'dd' work, in regards to keeping the file system. Does it simply copy files and disregards the structure, or does it make an exact copy, DataPlow FS and all? Anyone else ran into this conundrum?
I have many files in a folder from which I need to extract some contents, these are basically text files wich have individual lines with (i.e)
name: john address: whatever phone: 123456
Some caveats
1. Sometimes a line might be missing.
name: johnn phone: 123456
2. Lines are not in the same line-numbers across the files I did try some things with awk based on google searches but I couldn't extract the data of each file into a single line (this is the ultimate goal):
john,whatever,123456
I don't have knowledge other than having put some bash scripts together for backup jobs, so I am open to install anything that could to pull this off.
Is there any way to untar and only extract those files that are above a certain date including directory structure??
I restored a backup on a play server but it was a few days old. However I have a tar archive of the entire structure that is more up to date and healthy so now I want to extract all files (including directory structure) based on a date filter on the files if possible?
I have tar files where I archive about 250 files, each about 80 Mb, without compression. In a few cases tar is only returning some of the files. For example, when doing an extract of the file using: tar -xvf 356.tar I got only 103 files, when it should return 255 files, but tar does not give me an error. Furthermore, the tar archive is 15.8 Gb while the extracted folder is just 6.4 Gb. The tar files were created using: tar -cvf 356.tar 356 where 356 is the name of the folder. All the steps where done in the same machines, under Ubuntu 6 and newer. Any ideas if there is a way to recover the files that are not being extracted?
im trying to reconstruct / extract a file that was too large to fit onto a floppy, used 7zip to create and split the file into multiple parts in tar.bzip format. this was done in windows. Then moved all the parts of the file to tiny linux on a really old laptop. no cd drive, no usb or network. so have to rely on floppy drive. i do know that reconstruction while extracting using commands is possible. but not working.tried tar -xMf file.tar.001 but nothing.
IS there a program in ubuntu that extracts individual songs from album and label and tags them. i have just got 200 GB of music from a friend, but the albums are 1 full Mp3 file so it just runs right thru each song non stop. no individual song files what program can i use to convert album into individual song files and keep them together in a album. i found one that runs on windoze, but was wantiing to know if ubuntu has one. the windoze one is called Alba Extractor
how to go about extracting the .inf, .sys. and any .bin files from a driver. I found the windows driver for my laptop on the HP website, but it is a .exe file. How do I find the correct file?I am using the Broadcom BCM4312 chipset, on Ubuntu Maverick 10.10 unity 32bit.
my problem is that I can't extract some files because of a problem with cyrillic fonts (at least, this is what I suspect). The point is that I download some kind of files, which are fully functional to extract and open in windows but in Ubuntu trying the same thing, ends with an error. Files are downloaded from a russian web, so I think the fonts are cyrillic, which is the root of the problem. In windows they are shown as some kind of rare ASCII string but it's OK to open them.
The problem when trying in Ubuntu is that the file is shown as a large string with question marks '?' replacing all non-latin characters and it may cause problem, as it is the wildcard shell character in Linux. Renaming the file seems as an impossible task too. I have to process many of these files. Does Anyone know how can I avoid windows to do this? I tried to install many packages for cyrillic/russian support through synaptic nut it didn't seem to work.
This is what I tried to do:>cd ~/Desktop >sudo sh ati-driver-installer-10-1-x86.x86_64.run --extract.After entering my password, this appeared: >sh: Can't open ati-driver-installer-10-1-x86.x86_64.run.What am I doing wrong? It apparently worked in this thread.
how do I extract a pgp file on Ubuntu? I have the file and a passphrase. On Windows I have Kleoptra so I can right click the file and click 'decrypt'. Is there an easy way to do this on Ubuntu?
I know a .bin file is an executable file type in linux. We have an error after installing it and it referes to a file name and a line number within the file. I'm trying to find out if the file is part of the .bin file but I need a way to see what's inside of it or extract it.
and I want to extract VAR15 from each line (which can be at any column unfortunately - columns separated with commas - csv file), or VAR15 together with LATn,LONn from each line. Is it possible to do it with awk, grep or something other in linux?
I am trying to untar a file taken from true 64 bit Unix server to RHEL 5 server and I am getting a following error-Archive contains obsolescent base-64 headers.What can be done to extract the contains of those tar file.