I am trying to split a PDF into component pages that are of equal quality to the original. If I use display mypdf.pdf, then the PDF is split into pages of acceptable quality. The problem here is that I have to save each page individually. If, however, I use convert mypdf.pdf mypdf.bmp, I get the individual pages of the PDF in .BMP format (which is fine, but not exactly what I want), but the quality is substantially less than the original. I've tried dozens of combinations of commands to try to increase this quality, but to no avail.
Even if I do convert mypdf.pdf mypdfagain.pdf, there is a big loss of quality. Anyone familiar with splitting a PDF into individual pages without suffering a loss in quality? Ideally, I would just save all the "scenes/frames" from display, but that feature unfortunately does not exist (though I may endeavor myself to add it if no formal solution exists).
NOTE: I think part of my problem might be: by using identify mypdf.pdf I can see that the resolution is specified, and when I convert it the resolution is much lower. This could be a source of quality loss, but I'm not familiar enough with image conversion to say that for sure. Whatever this command does, it removes the extra layer or whatever it is that prevents OCR from succeeding. I'd really like to understand that technology.. What is it about a PDF that allows an individual to embed some meta-data into every page of the PDF so that the only thing seen, say, through OCR, or a text search function, is the embedded text?
How do I make imagemagick convert things one at a time? I know how to make imaggemagick convert all pages of a pdf into individual jpeg files by storing it all in the RAM at once but if there isn't enough RAM, this will use swap like crazy and be suuuuuuuper slow.
Here is the command for the "crazy swap-using" way:
I'm trying to use convert, I have installed the imagemagick. I use this line:convert *.jpg test.pdf but I'm only able to convert to pdf 1 single jpg file, not multiple files at once. When there's more than one file, I get the following error: Segmentation fault
I have been using the terminal command "convert" to make PDFs and I noticed that some of the pictures are out of order. The script I wrote is thus:
Quote:
#!/bin/bash cd /home/zach/Documents/PDFing convert *.jpg -resize 50% *.png
[code]...
During the stage where all the images are .png, they are still ordered properly in the folder. However upon converting to a PDF, some of them are in the wrong places. They are all uniformly named with numbers in the proper order, so that is not an issue.Is there a way that I can force "convert" to order strictly by name? The man pages were interesting, but did not help me in this situation.
Post Scriptum: I'm running 10.04 LTS Desktop Edition.
I have a large number of scanned JPG images that I would like to convert into a single PDF file -- each image will be on its own page, and I want the pages to be in a certain order.What is the command for ImageMagick to take a batch of JPGs and convert them to PDF, and order the pages in a certain way.
(by the way: I am working on writing a Gentoo ebuild for convert2pdf at the moment, but until I get that worked out, I'd like to figure out how to do this using Imagemagick, which is already in the portage tree, and already installed)
I keep trying to convert a bunch of jpg files into pdf, but ImageMagick just seems to keep failing there. Well well, after three thousand fix and reinstall attempts (seriously, I've been trying to fix it for the last month or so), this is what I'm getting today:
I have on my windows machine several hundred files that are a format of .nc .ncs for a CNC machine. I need to convert them to txt which is something as easy as opening in notepad and then saving as .txt but there are so many that this kind of action would take way too long.
The reason I am writing the linuxquestions is because I would feel more comfortable in loading a live CD and using some sort of terminal command to do this than I would to download one of the many "freeware" type programs I have found for windows (even more so since I have had a root kit before and had to start all the way over to get rid of it).
I need to know:
1. Is this possible to do with the terminal without super advanced knowledge.
2. Can one please point me in the right direction; something to read or an example
To convert web-pages into PDF file in ubuntu 10.10 is very easy way. just press Ctrl and P and you will see a new box called Print. In that you will see Print to file in printer section. Now just press Print to file now you can see Name box and Output Formate first you have select (.) PDF and and now you move to Name place and give and name to that file Exp:- xxx.pdf and now go to save in folder: place just press it to change to save on desktop. And press Print box. After pressing print box it will convert the information in PDF formate In the second way you have to go file --> print --> it will open print box as mention in above you have to convert the information into PDF format In the third way you have to go file --> print preview --> it will open a new box called Ubuntu homepage| Ubuntu - Mozilla Fire fox in that you just press Print box and it will open print box and do same as mention in first part it will convert web page into PDF format.
I know Ghostscript can convert PDFs to JPGs, and in the case of a multi-page PDF, can rip each page to an individual JPG. But is it possible to have it rip them to one JPG, so that the pages are pasted below each other, e.g. the top half of the JPG is page 1, the bottom half is page 2? Or do I have to use another program (and can ImageMagick do this?) to combine the JPG pages into one image?
I am learning about linux memory and hugepages, and know that hugepages basically is just memory that manages memory. I thought I'd experiment with the subject, and wrote a very small C program [URL] that basically just eats 20 GB of memory. The idea was that I would use this small C program to see how big the page table would get when handling large areas of memory when I'm not using hugepages. After running the program on my RHEL 5 server I was expecting the PageTable to be huge, but found that it was only about 43 MB. The page size on my RHEL box i 4 kB. Why I'm not getting the major PageTable size issue I was expecting?
On computer, I have apache server, configured at port 80, running. In local network, web page is accessible by my local IP (192...). Although, when I want to get to the page from the internet, I don't know right ip address, since router's one should be local, and also, router has it's own web page running at port 80. So I guessed that my server IP would be modem's one, which I don't know. Sometime in past, I figured out address of modem too, but it has it's own application running. Also, router supports port forwarding (which I guess, is needed), but I don't know how to get using to it. what address from internet should I have? How should I determine it?
im running ubutu 9.10 with the standard nvida proprietry drivers as a media center. the output is over hdmi, through a a/v reciver and onto a lcd television. my proble mis that the output is too large for the screen even whan outputting at a smaller resoultion than the native resolution of the television. this means that the top, bottom, left and right of the output is chopped off. in windows i can adjust the size of the output useing some sliders to get the image to fit the screen. however these are only available in the windows drivers and i cannot find a substitute in the ubuntu ones. one way of doing it may be to output black bard at the sides of the screen but i have no idea how to do this.
I have a PDF File here with upright ("portrait" format) pages, and I want to convert it so that each page gets "split" in the middle and becomes two wide ("landscape" format) pages.
Something like "print multiple pages on one page", only reversed
Is there any way to do such a thing with CUPS or something?
Using a default terminal and bash, there is no functionality to search the standard output of commands.
One can gain such functionality using other tools, like emacs shell or screen, but I am wondering why such a useful feature is missing, I do remember a simple C-F used to work in terminals.
Is there a way to make the Gnome terminal app support output search? or is there a better terminal app that support searching output natively?
I have been having a lot of trouble lately with installing from CD/DVD. The DVD reader/writer on this laptop is new. Nevertheless, trying to install Ubuntu onto an exernal HD, I get 'input output error on sr0 logical block (a large number) After a long time the booting proceeds to a point, but I never get the actual installation started, and have to shut down manually.
The CD is fine, says the Ubuntu-checker. I just installed using my sons laptop, and there was no trouble. Question: does this indicate a motherboard failure? A memory block damaged? Do you know of a diagnostic tool I can use to check the reading of a CD/DVD?
I have a konica minolta biz hub C550. When I am printing 11x17 the output is keeping an oversize margin at the top of the page no matter what setting I try... I have looked at the image shifting options... and other things. No matter how I shift the copy the top is getting cut off.
Just wanted to see if any one had this experience? I just did a fresh install of the recent LTS and saw it made a whopping 8 1/2 gig swap on a new pc I have 3 gigs of ram in? Im a try to resize as I have already done my customizing to it.
Directory /path/to is mounted over NFS (same network). Some options of fstab that might be helpful:
Both rsize and wsize are 3kB hard,proto=tcp,timeo=600,retrans=2
Sometimes the imagemagick command fails and produces no file (/path to/photo_for_public might not emerge after command execution). Files might be from 100kB to 10MB. We usually re-run the command for files not generated. Sometimes it fails even twice, so we do it third time (hilariously funny, but even a third run can fail).
Are there any thoughts what may be wrong with my photo-processing setup?
while installing ImageMagick on my i686-pc-linux gnu, everything seems to go fine when i put ./configure --with-modules in action. however, after this, as soon as i execute "make"
All has been fine with my emails but today morning when i tried to log on I got an error message "You must be logged in to access this page, go to logon page" I dont understand why even when i supply my correct user name and password.