Ubuntu :: Convert A Bunch Of Jpg Files Into Pdf - ImageMagick No Decode Delegate
Nov 9, 2010
I keep trying to convert a bunch of jpg files into pdf, but ImageMagick just seems to keep failing there. Well well, after three thousand fix and reinstall attempts (seriously, I've been trying to fix it for the last month or so), this is what I'm getting today:
I'm trying to use convert, I have installed the imagemagick. I use this line:convert *.jpg test.pdf but I'm only able to convert to pdf 1 single jpg file, not multiple files at once. When there's more than one file, I get the following error: Segmentation fault
I have Ubuntu 9.04 and just installed Sound Converter. I am trying to convert a bunch of .ogg files to mp3 to play on my iPod and it's not working so well. In the Sound Converter options I have is set to convert to high quality mp3. I choose the folder that the files are in and after a moment (slow laptop) Sound Converter populates, I hit 'convert' and it shows that the conversion completes in two seconds. All that it did was create the new folder structure of artist/album but there is nothing in there. Not sure what I am missing. I used Sound Converter before and it worked fine.
I have a large number of scanned JPG images that I would like to convert into a single PDF file -- each image will be on its own page, and I want the pages to be in a certain order.What is the command for ImageMagick to take a batch of JPGs and convert them to PDF, and order the pages in a certain way.
(by the way: I am working on writing a Gentoo ebuild for convert2pdf at the moment, but until I get that worked out, I'd like to figure out how to do this using Imagemagick, which is already in the portage tree, and already installed)
I'm not asking for help here, just documenting something I just discovered. Yesterday I wanted to batch-convert a bunch of old wma files to ogg vorbis. Not wanting to go through intermediate wav files, I tried to use ffmpeg to do it in one go. I first tried using the following command (in a loop, which I won't print here).
Code: ffmpeg -i $file -f ogg -acodec vorbis -ab 192k outputdir/$file "vorbis" turns out to be the built-in libavc implementation of the codec. In the process I discovered that the -ab value is always ignored. No matter what value you put, the output is always the default 64k (average, but of course it's vbr). You can however use the poorly-documented -aq option to set the audio quality used. The values don't correspond to the oggenc values though, being a number ranging from 10-100 (or more, I don't know what the maximum is). It's not exactly clear what number corresponds to what average bitrate, so you have to experiment. ~30 seems to give you an average-rate file, while anything above 60 is probably overkill.
Switching to the external libvorbis gave me more flexibility, although at a cost of much longer encoding times (note that ffmpeg must have been compiled with libvorbis support first).
I could use both -ab and -aq (with the numbers corresponding to the oggenc values), with no problems. ffmpeg does display some wrong values in it's output text, however. In addition, there's one more difference. The vorbis (libavc) codec provides an entry in the header of the ogg container reporting the average bitrate, but it doesn't appear to provide a similar bitrate header in the vorbis stream itself. Some programs may not report the bitrate value because of this.
libvorbis provides both headers, avoiding that problem. So to summarize, libvorbis appears to be a better codec choice than vorbis.
How do I make imagemagick convert things one at a time? I know how to make imaggemagick convert all pages of a pdf into individual jpeg files by storing it all in the RAM at once but if there isn't enough RAM, this will use swap like crazy and be suuuuuuuper slow.
Here is the command for the "crazy swap-using" way:
I am trying to split a PDF into component pages that are of equal quality to the original. If I use display mypdf.pdf, then the PDF is split into pages of acceptable quality. The problem here is that I have to save each page individually. If, however, I use convert mypdf.pdf mypdf.bmp, I get the individual pages of the PDF in .BMP format (which is fine, but not exactly what I want), but the quality is substantially less than the original. I've tried dozens of combinations of commands to try to increase this quality, but to no avail.
Even if I do convert mypdf.pdf mypdfagain.pdf, there is a big loss of quality. Anyone familiar with splitting a PDF into individual pages without suffering a loss in quality? Ideally, I would just save all the "scenes/frames" from display, but that feature unfortunately does not exist (though I may endeavor myself to add it if no formal solution exists).
NOTE: I think part of my problem might be: by using identify mypdf.pdf I can see that the resolution is specified, and when I convert it the resolution is much lower. This could be a source of quality loss, but I'm not familiar enough with image conversion to say that for sure. Whatever this command does, it removes the extra layer or whatever it is that prevents OCR from succeeding. I'd really like to understand that technology.. What is it about a PDF that allows an individual to embed some meta-data into every page of the PDF so that the only thing seen, say, through OCR, or a text search function, is the embedded text?
I have been playing with JWM source and found this cool tutorial at Debian Forums about how its easier to generate a .deb vs installing from source in the traditional manner (./configure, make,etc)[URL] My problem is that when doing the command
dpkg-buildpackage -rfakeroot -us -uc
it starts over, destroys the previous jwm stuff, including my custom files and generates a .deb so, how do I stop it from "cleaning" when i run the above command?
I keep a backup of a bunch of files on a flash drive, so that whenever I change distributions I can just restore all my Android stuff (saves on re-downloading everything). One of these is the Android SDK.
In my ~/.bashrc I add the paths to some executables in the SDK, only if the directory exists, and only if the path is not already in $PATH. For the Android NDK this works fine, but for the SDK I get this:
Code: snfo@snfo:~$ adb devices bash: /home/snfo/Android/sdk/platform-tools/adb: No such file or directory snfo@snfo:~$ ls -F /home/snfo/Android/sdk/platform-tools/adb /home/snfo/Android/sdk/platform-tools/adb*
Everything else is fine though, just that one path is causing trouble.
Now, I've saw something similar to this before whenever you move an executable from one place to another. If you don't re-source your bash config it will continue to keep looking wherever it used to be located. But I've never moved these files.
Basically I need to rename a bunch of .doc files using the for-structure and the mv command (w/ wildcards) in bash. I guess this would be a bit easier if I'd use the rename command, but since this is a school assignment of sorts I need to use for & mv. The .doc files are named "1filename.doc", "2filename.doc" etc. And I've got to rename them to "aaa_1filename.doc", "aaa_2filename.doc", "aaa_3filename.doc" and so on. Tried to dabble quite a bit with the for and mv commands, basically just got a bunch of errors. Every damn time. For 2 hours. The most common error was "mv: missing destination file operand after ..."
I have an rsnapshot backup that I need to move off of a corrupt Linux file system. I need to preserve the internal hardlinks. I've tried rsync -H and using a newer rsync and neither preserve the hardlinks on OS X.
I tried to get rsync -H working and I've isolated it to the file system mounted. I can preserve hard links copying locally (HFS to HFS) but it doesn't preserve when I try to rsync off of a SMB file system mount or AFP file system mount. Is there some mount option solution to getting OS X rsync to obey -H?
is there a simple shell script that would recurse all /home/xxx/public_html directories, and then yank this line (it will always be exactly the same) and better yet, for future, is there any way I can REPACE that line with another..
I have a lot of .flac files downloaded from several sites. Most of them come with a .cue file, and the .jpg with the cover, etc. It seems it is the intention of the uploader that one rebuilds the original CDDA. However, if I had a stand-alone CD/DVD player with flac I would hardly see the point of converting the flac to cdda. Furthermore, I could even play the flacs with a software player although, in this case, the audio quality would not be so good due to the noise picked up by the signal from the PC digital circuits.
My aim was to build a VERY simple smtp proxy under debian to handle mail from a port (51234) and forward it to the standard 25 port. I compile and install a "delegate" witch can handle easily that. It's working very well like that : delegated SERVER="smtp://anotherSmtpServer:25" -P51234
The strange thing is, it's working on my virtual test machine and on the dedicated server in local but I can't manage to use it trought internet. I test it like that. telnet [mySrv] 51234 Of course, no firewal, no deny host, no ined/xined, the service delegated is listening on the right port ... 2 clues : The port is answering trought internet with nmap as "51234/tcp open tcpwrapped" have a look at the tcpdump following :
On a bunch of servers, we use integrity checking tool on a large amount of data. Therefore, we compute SHA1 hash from a lot of files everyday. The CPU is a "generic" one, enough powerful for normal use, but get's totally saturated when integrity check begins. Is there any kind of server using a crypto-processor so as to delegate crypto functions to this dedicated processor?
I have been using Eclipse to program in Java for a long time, and have been using Jigloo for most of my GUI design. Recently I decided to try out the C/C++ plugin. I tried installing it, but it did not work. After many attempts, I finally gave up. However, after uninstalling it I found that Jigloo had stopped working. Any other plugins I try to install now also don't work. I have tried uninstalling and reinstalling plugins, and reinstalling eclipse, but nothing I have done fixes it. Although eclipse says that the software has been installed, nothing shows up in the features list or the plugins list.
Another problem: When I install eclipse from the repositories, it doesn't install a bunch of libraries and eclipse cannot recognize standard java files.
I need to search a bunch of files in a specific folder for a specific number and add all the numbers together to a total sum. I use Rsync everyday, everytime I run rsync i get a logfile (rsync output) witch contains the textstring "Total bytes sent: xxxxxx".
The "xxxxx" can vary in lenght. I need to extract the "xxxxxx" from each file and add the numbers together to a total size over a week or a month. Is this possible? And I wish to only use bash. One way of doing stuff at a time my friends .
Why does centos-release-notes have any dependancies ?
I can see no reason why centos-relese-notes should depend on centos-release when I try "yum erase centos-release-notes" I am told there are 72 dependancies including completely irrelevant packages such as tar all because of this unnecessary dependancy.
The centos-release-notes package simply contains a bunch of text/html files that nothing else relies upon, so why the dependancies ?
How do you convert Open Office (ODT) documents to Text files? I have made a report using libre office. Now I wish to continue editing the document using lyx (latex front end). So the ODT file needs to be saved as some .tex file.
I don't see an option to do this in File menu (export/save as). So is there any other plugin to do this?
i have a large directory of .bsp files that i would like to convert .bz2 archives. I've been searching for some time and all i can find is the obvious compress multiple files into one large archive. If anyone knows how to convert each file individually, while retaining the original file name (testmap.bsp would be archived as testmap.bsp.bz2)
In my application I came across a new requirement where I have to convert RTF files to Word and PDF formatted files. I am searching for an API using which my java application can able to convert the above specified formats.