Slackware :: Limit For Maximum Amount Of Locked Memory?
May 9, 2011
When I run ardour sound editing I get this message , but it starts ok Your system has a limit for maximum amount of locked memory! This might cause Ardour to run out of memory before your system runs out of memory. You can view the memory limit with 'ulimit -l', and it is normally controlled by /etc/security/limits.conf
bash-4.1$ ulimit -l
64
my limits.conf is like this
audio - rtprio 99
@audio - memlock 250000
[Code]...
View 2 Replies
ADVERTISEMENT
Jul 24, 2009
I have a few multi-user servers in an academic laboratory. I am having a problem with some users maxing out the available RAM, causing such sever slowdowns the machine essentially crashes. My servers are Dell Power Edge's running Ubuntu 8.10 Server Edition (Not my choice). I would like to set a maximum limit on the amount of ram a user can utilize. This morning I experimented with setting limits via /etc/security/limits.conf and using ulimit. Neither of them prevented my test program, a simple infinite loop of mallocs, from crashing the server.
View 7 Replies
View Related
May 17, 2011
My problem is I installed Zone Minder for camera security and I'm testing it on my laptop with the built in webcam and everything seems to work perfectly except when I try to view the live feed from the camera, it's just a black box. No video.
I checked this website and it's exactly the problem I'm having with a fix for it but his fix doesn't work. He says to type:
Code:
user@ubuntu:~$ sudo echo "256000000" > /proc/sys/kernel/shmmax
user@ubuntu:~$ sudo service apache2 restart
user@ubuntu:~$ sudo service zoneminder restart
But when I do the first line, I get this:
Code:
deemar@Clementine:~$ sudo echo "256000000" > /proc/sys/kernel/shmmax
bash: /proc/sys/kernel/shmmax: Permission denied
View 2 Replies
View Related
Nov 15, 2010
I'm running Ardour, and when I launch it I get the following:
Code:
WARNING: Your system has a limit for maximum amount of locked memory!
This might cause Ardour to run out of memory before your system runs
[code]...
View 8 Replies
View Related
Feb 14, 2009
I will be upgrading my RAM soon. What is the maximum amount of ram a 32bit OS can see?
View 14 Replies
View Related
Jul 15, 2011
Once turned on and turned off the desktop effects was a change in the settings windows, always open windows in the maximum amount for all programs. How to Recover the initial setting?
View 5 Replies
View Related
May 12, 2010
I have a VPS server with 512 MB memory. The php.ini is set so script memory limit = 16 MB. However, I have noticed in my top report, instances like the following:
Quote:
5484 coldclim 25 0 46476 32m 5920 R 0.0 6.4 0:00.93 php
The bold number of 6.4 is the % of sever memory this process is using. 6.4 % of 512 MB of memory is about 32 MB of memory, so it appears that this isn't being limited by php.ini. Am I correct? This leads to the next question: Is there some way to limit the amount of memory a single suphp process can use? (Basically, something like the setting in php.ini which limits suphp processes in the same way.)
View 2 Replies
View Related
Aug 17, 2010
I want to use the tc rules for bandwidth control in my lan.i have a linux router(traffic shaper).how i can limit the amount of bandwidth every user has access to per day? For example, any user can't download more than 2 gb per day(or per month).
The information about lan's users(such as a group type,userid,etc.but not any thing about time limiting per user)is in ldap directory on ldap server.the linux server uses ldap server for authentication users when the user login.
View 6 Replies
View Related
Jun 30, 2010
I was wondering if there was a Windows or Ubuntu way to limit the amount of data that is able to be sent over the internet between certain times, eg. Between the times of 7am and 7pm can only download 300 MB from the web, when this limit is reached the web is either disconnected or slowed down.
View 2 Replies
View Related
Jul 1, 2010
tell me the maximum limit of connections FTP server
View 2 Replies
View Related
May 4, 2011
I'm running nginx for static files and as a proxy server for a comet IM server on ubuntu Jaunty. On high load I'm hitting a limit of 1024 file descriptors. I've tried increasing this limit but still can't pass 1024. Does "more /proc/sys/fs/file-nr" gives me the global count of used file descriptors? Why do I see a maximum of 1024 open file descriptors in /proc/sys/fs/file-nr if this is the global count for the machine and each user should have at least 1024 allowed file descriptors by default? Is there a way to increase the limit while the server is running?
Some relevant info on my server:
sudo more /proc/sys/fs/file-nr
1024038001
sudo sysctl fs.file-max
fs.file-max = 38001
sudo nano /etc/security/limits.conf
...
* hard nofile 30000
* soft nofile 30000
I also added this to /usr/local/nginx/conf/nginx.conf:
worker_rlimit_nofile 10240;
Uncommented the following line in /etc/pam.d/su:
session required pam_limits.so
View 1 Replies
View Related
May 14, 2009
On a new box with Debian Lenny, I have 2x2 gig of ddr2 but the karamba applet show only 3291mb. And dmesg show this:
Code:
dmesg | grep Memory
[0.004000] Memory: 3362976k/4194304k available (1769k kernel code, 43092k reserved, 752k data, 244k init, 2489792k highmem)
View 2 Replies
View Related
Jun 4, 2010
as you should see, top is indicating 3.544.864kb (3.5Gb) of memory used while gnome system monitor only 609Mb. What's wrong here? (I am pretty sure Gnome SM is right. Top is updating every sec.)
View 1 Replies
View Related
Nov 8, 2010
I need to allocate a % of the total system memory for a buffer but what is the best method to determine how much memory is in the system? So far the only way I have found is to get the pages of memory:
Code:
long sysconf(_SC_PHYS_PAGES)
Is that the only option?
View 2 Replies
View Related
Mar 3, 2009
I would like to know if there is a linux command to verify the amount of memory used by a program. The programs I am using were compiled with gfortran.
View 2 Replies
View Related
Aug 25, 2010
Is there an easy way in C to get the total amount of allocated SHM memory on a machine?
View 7 Replies
View Related
Jul 15, 2010
I want to calculate the total amount of memory used by the system, Someone please confirm if the below formula is correct.Quote:
((memUsed-mebBuffers-memCached)*100)/memUsed => totalMemoryUsed%
Code:
free -o
total used free shared buffers cached
Mem: 8167004 8115292 51712 0 87680 1444676
Swap: 8388584 3252996 5135588
((8115292-87680-1444676)*100/8115292) => 81%
As per the above calculation 81% of memory is used.Is this correct? and if so Am I running out of memory?what is the limit in % that I should maintain for a better performance?
View 3 Replies
View Related
Mar 26, 2010
I am using a machine I SSH into to do some processing over very large files, but I am now unable to access it through ssh.The shell I had opened is not responsive, and when I try to connect again, it hangs even before asking for the password.I suspect this to be a memory problem as a couple of processes need to load a lot of the data to process it.Is there an emergency way of going back into this machine, or killing some processes to free the memory up ? I tried ssh ... 'kill -9 pid' but it is not even able to do this.Do you reckon the processing is still happening or is the machine completely idle ? How would you advise me to go back to it ? It's for my final year project, and I'm starting to stress !
View 2 Replies
View Related
Mar 24, 2010
I am trying to understand a large amount of allocated memory that seems not to be accounted for on my system.I'll say up front that I am discussing memory usage without cache and buffers, 'cause I know that misunderstanding comes up a lot.I am in a KDE 4.3 desktop (Kubuntu 9.10), using a number of java apps like Eclipse that tend to eat up a lot of memory.after a few days, even if I quit most apps, 1 gb of ram remains allocated (out of 2 gb).this appeared excessive, and I took the time to add up all values of the RES column in htop (for all users).the result was about 1/2 gb.am I trying to match the wrong values?or could some memory be allocated and not show up in the process list?this is the output of free
Code:
total used free shared buffers cached
Mem: 2055456 1940264 115192 0 123864 702900
[code]...
View 6 Replies
View Related
Apr 4, 2011
I am trying to write a script to calculate the total amount of installed memory to use during an anaconda kickscript, so the swap file is created at 2 x the installed memory. I so far have the amount of installed RAM DIMMS but need a way to total them up and produce a varible I can use in the pre section of the install.
Code:
dmidecode -t 17 | grep Size | awk ' { print $2 }'
Output:
2048
2048
Note: on some servers there could be from 1 DIMM up to 16 DIMMS installed so the script needs to be able to handle this. I also can not use bc as it does not exist during the install stage. I am guessing I need a while loop to do this and use expr but do not know where to start for this logic.
View 3 Replies
View Related
Jul 17, 2011
How do I change the size of the available shared memory on Linux?evidently 4GB is not enough for what I am doing (I need to load a lot of data into shared memory - my machine got 8GB of RAM).
View 2 Replies
View Related
Mar 5, 2010
First time here at LQ, so if the post is in the wrong place, this is a bit programming related. I've been trying to find the maximum resident memory of a process in a memory constraint situation.
Instead of using top, or ps, which gives me real-time snapshot at the memory usage of a process, is there a way to determine the maximum resident memory used by a process? /usr/bin/time seem to provide this functionality within the format string
[Code]...
The minor page-fault is suspicious, does that mean memory is being used.. but not recorded? if so, how can I know how much memory is being mapped?
View 3 Replies
View Related
Apr 24, 2010
I've written a program for a class that my professor will be testing in various low memory environments to see how it behaves when the program runs out of memory. Is there a way I can simulate the execution in a low memory environment without creating a virtual machine?
View 1 Replies
View Related
Jun 26, 2010
I have a 32 bit Ubuntu installed and my Laptop has 4GB RAM, but only 3GB is considered by Linux. My question is: what is the reason for the upper limit on physical memory ?
Code:
dmesg | grep Memory [0.000000] Memory: 3052428k/3112960k available (4673k kernel code, 56364k reserved, 2121k data, 656k init, 2200904k highmem) I am familiar with the virtual memory concept where linux splits upper 1GB for kernel and lower 3GB for user processes. In total, linux 32bit can address 4GB virtual addresses. Does this meant that 1GB of physical memory is already mapped to 1GB of kernel space and Linux only shows the remaining 3GB physical memory left for the user in the above command.
I did some searching on the internet and found some articles related to this, but it only confused me further since some articles suggest 4GB is the upper limit with mentioning whether it's virtual or physical memory, some bring in the concept of PAE, etc. I'm relative new to Linux's memory management, so it'd be really helpful if someone could answer this.
View 4 Replies
View Related
Aug 13, 2011
I was looking into using control groups to limit the memory usage of each user on my CentOS system. I was told that this required me to recompile the kernel to have cgroup support. Is this true? Or is there a kernel module that will allow cgroups to work for users and groups on the system without kernel re-compile? Or, is there another way to limit the users memory usage? I have tried ulimit and it doesn't seem to work right.
I ask since this setup will be on a VPS system, that means to re-compile the kernel I need to use Xen instead of OpenVZ. Plus I have never in my life re-compiled the kernel, least of all with different modules ha ha ha so I would have to pay my NOC to do it. So if I don't HAVE to recompile the kernel to get cgroup support.
View 2 Replies
View Related
Dec 22, 2014
How to limit access flash memory (read / write) for users on debian?
View 2 Replies
View Related
Jun 25, 2010
I presently have Fedora 12 running on a Dell Optiplex with a Pentium 4 CPU. I want to buy a new computer, probably from Dell, I don't completely understand the options I have available for which Fedora 13 will run.I more or less understand what 64 bit means since I also have an HP with an AMD 64 bit processor. But are there other options which will allow me to exceed the i386 memory limit? How about dual processors?
I used to understand CPU architecture pretty well, but i've lost track of more recent developments. Can anyone recommend a primer of CPU types, including assoicated memory limits? Similarly, I would like to be sure that Fedora 13 will run on whatever machine I decide to get. So a list of available processors, which are equivalent to which, and which Fedora 13 has been tested on,
View 7 Replies
View Related
May 28, 2010
I have an Ubuntu server running in our small office. Among its many duties is report generation. It uses PHP and DOMPDF (a PHP library for converting HTML/CSS to PDFs for printing). PHP's default memory limit of 32MB is not even close to being enough to pull large amounts of data from the database and generate images/tables/PDFs with that data.
I increased the memory limit to 64MB and that is adequate for reports under 3 pages or so (varies based on table complexity, images, etc). If any user tries to generate a report longer than that, PHP just throws a "out of memory" error and doesn't make the report.
My question is: what are the possible consequences of increasing the memory limit yet again to 128MB or maybe even higher? The server isn't terribly powerful. It has 2GB RAM and 4GB swap space. I know that isn't much but this is a small office and at most I can only see two or three people trying to run reports at the same time. As for security, apache is currently only serving pages in the local network, but sometime within the next year I'll probably have it hosting a public website (currently using a hosting service). Is a high memory limit a potential security risk when exposed to the internet?
EDIT: Sorry, PHP's default memory limit is 16 not 32 as I said. Question still stands, however
View 5 Replies
View Related
Sep 2, 2010
I'm having a problem with setrlimit() under linux.
If i used setrlimit(RLIMIT_AS) to set a hard ceiling of virtual memory usage first, then request memory more than that, shouldn't i receive a signal like SIGSEGV?
First i tried the command ulimit in bash, which acted as if i called setrlimit(). i tried 3 programs that overflowed the memory limit i set. i though all the programs would be terminated by a SIGSEGV. but the pascal program received SIGKILL; the C++ program got SIGABRT› the C program, SIGSEGV.
I though there maybe something different between setrlimit() and ulimit, so i wrote a program in C++, fork() first, setrlimit() then execv() in the child, and wait(&status) in the parent. but i got the same result.
I was wondering why these could happen, and could anyone tell me how could i deal with them? i mean, how can i judge that the program exited abnormally because it exceed memory limit?
View 2 Replies
View Related
Feb 23, 2011
Fresh install of Debian 6.0 (squeeze) and the memory shows 3.8 GB. I have 6 sticks of 2 gb on the board.
Pertinent dmesg lines: AMI BIOS detected: BIOS may corrupt low RAM, working around it.
3083MB HIGHMEM available.
883MB LOWMEM available.
Bios tests 12 gb of ram fine.
View 4 Replies
View Related