Software :: Pdftk 'File Size Limit Exceeded' Error?

Apr 12, 2010

I downloaded pdftk 1.41 fromand installed on Red Hat Enterprise Linux 4, 32 bitI am primarily using this utility to uncompress pdf files to remove the 'Flate' compressionIt works good with small pdfsHowever, when i use to uncompress pdf files of size 35MB or more, the uncompressed output file grows up to 2GB and then the uncompression fails with error:"File size limit exceeded"I can concatenate two files with output file size upto 3GB in size, so 2GB is not the limitation at the linux level

View 1 Replies


ADVERTISEMENT

Software :: File Size Limit Exceeded Error?

May 7, 2010

I'm trying to copy a 7.8GB tar.gz file to an external hard drive via command line. It gets to an even 4GB and stops, and gives an error that says "file size limit exceeded." I edited some file at /etc/security/limits.conf to look like: "root hard fsize 10024000" but that didn't do anything at all. Yes, I am copying this as root.

View 9 Replies View Related

Ubuntu :: Pdftk Output File With Pages Extracts Same Size As Input File?

Apr 22, 2011

I am curious if perhaps I am doing something wrong extracting pages from a pdf doc using pdftk and creating a new file. I am only extracting the odd pages from the file and outputting them to a new file that is now only 20 pages instead of the input's 40 pages, yet the new output file is still 1.4Mb in size, the same as the original.

It seems strange to extract only half the pages of a large document and end up with a result that is the same size. how to streamline the resulting pdf's using pdftk?

BTW this is the command I am using, in case perhaps I am missing an option to optimize file size or something:

Code:
pdftk A=ch15.pdf cat A1-40odd output odd.pdf

View 1 Replies View Related

Debian Installation :: Dpkg-scanpackages Fails When Package Limit Exceeded

Nov 2, 2015

After receiving no response either here or on IRC, I copied 80 package files to a temporary directory and ran dpkg-scanpackages . /dev/null > Packages with the expected result. The curious part is the delay when output redirection is not used: nothing appears until the script completes, when the result is dumped to the screen. It therefore appears that there is an upper limit to the number of packages that the script can handle, somewhere between 80 and 42,474.

Is this an undocumented feature, or just a peculiarity of my system?

I'm new to Debian and wanting to set up a local repository on my work drive. After following instructions online and copying all packages (~43,000) from the DVD set into /work/Debian/8.2/packages/ I ran dpkg-scan-packages as instructed:

Code: Select all# dpkg-scan-packages . /dev/null | gzip -9c > Packages.gz

This produced an empty file. I then ran dpkg-scanpackages with no output redirection expecting to see a flood of text on the screen, but all I got were error messages suggesting that it can see the .deb packages but is not parsing them:

Code: Select allroot@qbx:~# dpkg-scanpackages /work/Debian/8.2/packages
dpkg-deb: error: invalid character ' ' in archive '/work/Debian/8.2/packages/libshhopt1_1.1.7-3_i386.deb' member 'debian-binary' size
dpkg-scanpackages: error: couldn't parse control information from /work/Debian/8.2/packages/libshhopt1_1.1.7-3_i386.deb

[Code] ....

This all seems in accord with the man page, and it's so simple I'm wondering what I'm missing.

View 1 Replies View Related

Programming :: Why Different Signals Send To Programs That Exceeded Memory Limit Set By Setrlimit()?

Sep 2, 2010

I'm having a problem with setrlimit() under linux.

If i used setrlimit(RLIMIT_AS) to set a hard ceiling of virtual memory usage first, then request memory more than that, shouldn't i receive a signal like SIGSEGV?

First i tried the command ulimit in bash, which acted as if i called setrlimit(). i tried 3 programs that overflowed the memory limit i set. i though all the programs would be terminated by a SIGSEGV. but the pascal program received SIGKILL; the C++ program got SIGABRT› the C program, SIGSEGV.

I though there maybe something different between setrlimit() and ulimit, so i wrote a program in C++, fork() first, setrlimit() then execv() in the child, and wait(&status) in the parent. but i got the same result.

I was wondering why these could happen, and could anyone tell me how could i deal with them? i mean, how can i judge that the program exited abnormally because it exceed memory limit?

View 2 Replies View Related

General :: Limit On File Size - Doesn't Allow User To Create File Which Are Greater Than 100Kb

Jun 13, 2011

I was just testing specifying limit on file size to a user and have added the following to /etc/security/limits.conf bob soft fsize 100 This basically should have said not to allow bob to create anyfile greater than 100Kb in size.

But the interesting thing is, if bob already has any file which is greater than 100Kb in size, it even doesn't allow to log him into the system both from console and SSH. Also nothing is logged in logs.. How do I configure it so that, bob can login to the system even though he has any file greater than 100Kb (but doesn't allow him to create file which are greater than 100Kb) ??

View 3 Replies View Related

Ubuntu :: Limit Directory Size \ Set A Size Limit On?

Mar 2, 2011

I have 2 directories in my home folder that I would like to set a size limit on. The directories are ~/backup and ~/temp. Is there an easy way to limit the size of a directory without having to make partitions?

View 4 Replies View Related

Ubuntu One :: File Size Limit For Upload?

May 11, 2010

I have a large file (deflated size: 602191947)that is not saved in my Ubuntu One account. On sync'ing the file is being uploaded, and eventually reaches 602191947 - and then nothing more happens to this file - but sync'ing the following files in the queue goes on with success. I have tried manual upload with the same result. The file is still being marked as 'uploading' even after several tries and log ins/log outs, and reboots. So I was just wondering whether there is a file size limit - can't seem to find information regarding this.

View 5 Replies View Related

Ubuntu :: Swap File Size Limit?

Oct 20, 2010

a possibly preposterous question. I am aware that you can designate a swap file or swap partition on your hard drive that linux uses as "memory". Suggested sizes for the swap file that I've seen range up to about 1024MB. Is there a limit to the swap file size that you can set?Basically I am running a perl script that processes a massive B) file (DNA sequence data), etc, and requires around 48 GB of memory to run, maybe a bit less. So, would it be possible to set a swap file to a massive, ridiculous size (~60GB oratever) and successfully run such a script on a desktop?Yes, I am aware that it would massively ow down the process. The thing is, if the perl script normally completes in about half an hour, and I can get it working on a desktop, I don't mind if it takes days or weeks to complete. I really don't. That's because it takes days or weeks to get access to a computer with the required grunt to do it.So, is this a stupid idea? Is it even possible? If so, given a perl script that normally completes in a half hour on a 48G system, if you do this, would it take days? weeks? decades

View 7 Replies View Related

Ubuntu :: File Size Limit For Gedit?

Nov 5, 2010

I've noticed that for files longer than about 8000 lines that gedit has problems opening the file. Was gedit not designed for long files or is there another problem? The same thing also happens on complicated html files. So I hope there is a way to fix this.

View 4 Replies View Related

Software :: Limit Log File Size With Syslog-ng?

Jan 4, 2010

I have a self-made application running on a small embedded Linux device (which should not matter) using syslog to output some error, warning or debug logs.There is a "better" syslog daemon installed, called syslog-ng, which have some more features,t I miss a very important one:How to limit the size of the logfiles to some dedicated megabytes. I was able to create rotating logfiles with the configuration in syslog-ng.conf:

Code:
destination testlog {
file("/var/log/test/log-$S_WEEKDAY"

[code]...

View 2 Replies View Related

Fedora :: Circumvent Fat32 4Gb File Size Limit?

Dec 16, 2010

I have a single 6.2Gb file that needs to go on a fat32 format hdd, does anyone know of a way to split the file so it will fit.

View 2 Replies View Related

Ubuntu Multimedia :: Recordmydesktop File Size Limit?

Apr 21, 2010

Does Recordmydesktop have a file size limit? I'm considering using the Zero compression setting to keep CPU usage down, but I don't want to run up against a 2GB or 4GB file size limit. While I know some filesystems impose this limit, most screen recorders I've used have a 2GB or 4GB limit when recording, regardless of the filesystem.Is this an issue with Recordmydesktop

View 1 Replies View Related

Programming :: Soft Limit For Core File Size?

Jun 22, 2011

Using getrlimit I am setting the core file size to be RLIM_INFINITY. But still the core file is not being generated,although in /var/log/messages it says a core is being generated

View 3 Replies View Related

Ubuntu Servers :: Limit Size Of Log File When Redirecting Stdout?

Jun 7, 2010

I have a command line server that logs to stdout, which I start along the lines of ./server > log.txt

What I want to do is limit the size of log.txt, without modifying the server.

I am assuming there must be some kind of tool already that lets me do this, something like where I can pass in my server, the output file and a size limit? If so, can anyone enlighten me?

View 3 Replies View Related

General :: Apache Access Log File Recommended Size Limit

Jan 17, 2011

Based on some web reference, apache access log file size increased more than 2Gb. It will effect apache performances. Is this correct?

View 4 Replies View Related

OpenSUSE :: Nothing Provides Pdftk Needed By Pdftk-qgui-0.1.9.2-0.pm.1.1.x86_86?

Aug 16, 2010

pdftk could be installed in the Suse 11.0 version, and worked very well. I love it. But recently, after I install Suse11.3 on my Dell desktop (64 bits), everything works fine except the pdftk. When I am trying to install it from Yast, it warns:nothing provides pdftk needed by pdftk-qgui-0.1.9.2-0.pm.1.1.x86_86, and gives me two conflict resolutions: 1. do not install it2. break pdftk-qgui by ignoring some of its dependencies.Obviously, none of them are the correct answer .

View 2 Replies View Related

Server :: New Open File Size Limit Is Getting Reflected In The Specific User?

May 16, 2011

Last weekend i have increased the open file size (ulimit -n) for the application user id i have update the limits.conf file with necessary inputs restarted the service and the server as well, when i check the ulimit value for the specific user by switching user from other user it shows the new value (10240) but if i login directly using the application id the ulimit value shows as 1024 which one is the default one.

View 6 Replies View Related

Fedora :: Bash - Ulimit: Core File Size - Can't Modify Limit: Operation Not Permitted?

Dec 7, 2009

Fedora 12 gcc 4.4.1 I am doing some programming, and my program gave me a stack dump. However, there is no core file for me to examine.

So I did:
Code:
ulimit -c unlimited
and got this error message:

Code:
bash: ulimit: core file size: cannot modify limit: Operation not permitted I also tried setting ulimit to 50000 and still got the same error. The results of ulimit -a:

Code:
$ ulimit -a
core file size (blocks, -c) 0
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited

[Code]...

View 3 Replies View Related

Debian :: Input File For Pdftk?

Apr 21, 2011

This is a question for those that use pdftk. I've been using it for about the last year or so to combine manipulate PDFs and it is awesome. Until today, mostly what I've been using it for has been for less than five files. Today I have been working on a manual at work which consists of about 100 separate files with a total page count of 167. The pages are a mix of 8.5x11(inch) and 11x17(inch). The pages needed to be in a specific order, the only way I could think of to do this instead doing this all in the CLI, I created a text file which contained a list or sequence of the files. To make it easier for me to read and group, I first put each file on a separate line.

chapter1_separation_page.pdf
file1.pdf
file2.pdf

[code]....

View 10 Replies View Related

Debian Installation :: Postfix Fatal Error: Unable To Determine File Limit

May 11, 2011

I upgraded postfix to 2.8 today after the security bulletin from the list.

now I get the following error:

'postfix/local[2442]: fatal: unable to determine open file limit'

I can receive/send mail, it just wont let me deliver locally

uname output: 2.6.37-1-686 #1 SMP Tue Feb 15 18:21:50 UTC 2011 i686 GNU/Linux

View 2 Replies View Related

CentOS 5 Server :: KVM + Internal Error Bridge Generation Exceeded Max Id 256

Jul 18, 2010

I did a fresh installation of centos 5.5. I installed KVM. And configured a eth1 for KVM host network and i configured eth0 to be bridged.

When i do service libvirtd reload i get: internal error Bridge generation exceeded max id 256

View 5 Replies View Related

Software :: Filesystem Size Limit Is For 5.1?

Apr 2, 2010

We use VxVM and VxFS on HP-UX and Ie used them in the past on Solaris.ve found they are available as Storage Foundation v5.1 for 64 bit RHEL.(In fact there�s even a BASIC version for free on 2 processor systems). Previously wed run into a 2 TB limit for filesystems on the older versions we have on HP-UX. The data sheets at Symantec are pure marketing fluff. Does anyone know what the filesystem size limit is for 5.1 on Linux?

View 1 Replies View Related

Programming :: Why The Stack Size Limit

May 1, 2011

Why is it in Linux that there is a stack size set by default? And why is it so small? (My system is set to 8192 kbytes.) And why is there a default limit on the stack size when the max memory and virtual memory size are, by default, unlimited? (Aren't they both fed from the same place ultimately?)

Reason I ask: I want to use recursive functions in my programming a lot more. Problem is, if the language (or implementation) doesn't happen to support tail-call recursion, then I can be pretty well certain that the first huge problem that gets thrown at my function is going to kill my program because the stack size limit is going to be quickly reached. Obviously, I can change the stack size limit for my own computers, but it doesn't feel so great knowing that most of the people who copy and execute my code will have probably have overlooked this. Anyway, does anyone know: is this small default stack size limit just one of those historical artifacts, or is there some technical reason for it?

View 5 Replies View Related

OpenSUSE Hardware :: Limit The Pagecache Size In 11.4?

Jul 7, 2011

How do you put a limit in place for file caching in Suse 11.4?

My pc becomes usable on a regular basis with minimum cpu usage because I can't open new applications There are no error messages etc, the new apps just don't open.

free -m shows the vast majority of my memory is used by cache
free -m
total used free shared buffers cached
Mem: 3185 1048 2137 0 38 503 (this becomes max)
-/+ buffers/cache: 506 2679
Swap: 2055 0 2055

I've done ...
echo 3 > /proc/sys/vm/drop_caches

Sometimes an app will open after this but the system becomes unstable, locking up regularly.

I not sure why the default max is 100% file cache but I'd to put a sane file cache limit in place, like 40% or something. I've put limits in place in the past using a percentage and I've poked arround but i don't see the setting.

View 3 Replies View Related

Ubuntu :: Limit The Download Size Of Updates?

Jun 17, 2010

Is there some way to limit the download size of updates for ubuntu? At the moment, update manager shows that I have some 300 MB worth of downloads. I can't find any way to deselect many updates at once either.

View 3 Replies View Related

Ubuntu Servers :: 3.0 Tb Partition Size Limit?

Jun 19, 2010

I really don't understand what's happening.I make a 3.5tb RAID array in Disk Utility, yet it makes it so that one partition is 3tb and the other is 500 gigs free!Why is that? Ext4 can do huge partition sizes I thought.

View 1 Replies View Related

Ubuntu :: Mount & Limit Folders Size?

Oct 24, 2010

I have a 1TB external hard drive. I would like to create in it 10 folders:

Code:

I would then like to permanently mount each folder to its machine (I have 10 machines connected through a switch, so each machine will have a folder that is mounted to ONE of the 10 folders in the external hard drive).

My questions:
(1) Is this a good configuration? are there better ideas to give individual machines more space without replacing their hard drive?
(2) How do I limit each one of the folders ('folder1', 'folder2', ...., 'folder10') to a size of 100 [GB]? I don't want one folder (say, 'folder1') to grow in size and 'steal' the space designated to the other folders.

View 2 Replies View Related

Server :: Limit Print Job Size From Samba?

Mar 15, 2010

Does anyone know of a way of limiting a print-job size from samba?

I know how to limit a print job size form cups, and how to require x amount of free space before accepting a job. I've even dug up how to require x amount of free space for samba to accept a print job, but I can't see how to limit samba to only certain sized jobs.

Someone tried to print a >1G file to my print-server this morning, causing me to have a less relaxed Monday than I had hoped. Because it ran out of space before spooling, it was never limited by cups. Because I had to get rid of it ASAP so people could get work done, I have no idea who's it was, or where it came from. Scouring logs didn't give me any good leads either.

View 2 Replies View Related

Server :: Postfix Message Size Limit?

Sep 1, 2010

I have been trying to increase the message_size_limit on my Debian 2.4.26 box with postfix 2.3.8. For example, I set message_size_limit and mailbox_size_limit to 104857600 (100m) and restart postfix. Running postconf -n confirms that it has changed. However when I send a test message it kicks it back saying the message size limit is 16777216 (16m, which is, incidentally, the default value of the berkeley_db_create_buffer_size parameter)

View 10 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved