Ubuntu :: Understand A Large Amount Of Allocated Memory That Seems Not To Be Accounted For On System?
Mar 24, 2010
I am trying to understand a large amount of allocated memory that seems not to be accounted for on my system.I'll say up front that I am discussing memory usage without cache and buffers, 'cause I know that misunderstanding comes up a lot.I am in a KDE 4.3 desktop (Kubuntu 9.10), using a number of java apps like Eclipse that tend to eat up a lot of memory.after a few days, even if I quit most apps, 1 gb of ram remains allocated (out of 2 gb).this appeared excessive, and I took the time to add up all values of the RES column in htop (for all users).the result was about 1/2 gb.am I trying to match the wrong values?or could some memory be allocated and not show up in the process list?this is the output of free
Code:
total used free shared buffers cached
Mem: 2055456 1940264 115192 0 123864 702900
I need to allocate a % of the total system memory for a buffer but what is the best method to determine how much memory is in the system? So far the only way I have found is to get the pages of memory:
Code: long sysconf(_SC_PHYS_PAGES) Is that the only option?
As per the above calculation 81% of memory is used.Is this correct? and if so Am I running out of memory?what is the limit in % that I should maintain for a better performance?
I am still confused about the memory issues. i have allocated JAVA_OPTS='-Djava.awt.headless=true -Xms512M -Xmx512M -XX:MaxPermSize=512M' think that it means that i have allocate java to use only 512mb of my memory. but when i run top here is my output.
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 15932 root 23 0 1495m 1.2g 3732 S 12.0 36.0 121:52.98 java 5070 postgres 15 0 55392 39m 35m R 11.6 1.2 0:51.61 postmaster
I want to know if i can increase the memory space allocated for a process manually while the process is running ,,,, and if it is possible how i can do this .
In gparted I have the following stats for my /home drive
size: 824 gb used 75.51 gb unused 748.59 gb
Now when I view this in nautilus it shows something else: remaining free space as 709 gb. My question is what happened to the 40gbs? the 75.51gb are my files, but where did the 40gbs go to? Because 709 (total remaining) + 75 (my files) + 40 (mysteriously lost gbs) = 824gb. When I first made the partiton, it was a 824gb partition and ubuntu had automatically at that point reserved about 40gb for something. Does anyone know why Ubuntu reserved this space?
I use Audacious to play my music... But every time I want to add my music it crashes... even when I do it in batches. Is there somewhere I can find logging to see what the problem is?
On a new box with Debian Lenny, I have 2x2 gig of ddr2 but the karamba applet show only 3291mb. And dmesg show this: Code: dmesg | grep Memory [0.004000] Memory: 3362976k/4194304k available (1769k kernel code, 43092k reserved, 752k data, 244k init, 2489792k highmem)
as you should see, top is indicating 3.544.864kb (3.5Gb) of memory used while gnome system monitor only 609Mb. What's wrong here? (I am pretty sure Gnome SM is right. Top is updating every sec.)
I would like to know if there is a linux command to verify the amount of memory used by a program. The programs I am using were compiled with gfortran.
My problem is I installed Zone Minder for camera security and I'm testing it on my laptop with the built in webcam and everything seems to work perfectly except when I try to view the live feed from the camera, it's just a black box. No video.
I checked this website and it's exactly the problem I'm having with a fix for it but his fix doesn't work. He says to type:
Code: user@ubuntu:~$ sudo echo "256000000" > /proc/sys/kernel/shmmax user@ubuntu:~$ sudo service apache2 restart user@ubuntu:~$ sudo service zoneminder restart
When I run ardour sound editing I get this message , but it starts ok Your system has a limit for maximum amount of locked memory! This might cause Ardour to run out of memory before your system runs out of memory. You can view the memory limit with 'ulimit -l', and it is normally controlled by /etc/security/limits.conf
bash-4.1$ ulimit -l 64 my limits.conf is like this audio - rtprio 99 @audio - memlock 250000
I am trying to write a script to calculate the total amount of installed memory to use during an anaconda kickscript, so the swap file is created at 2 x the installed memory. I so far have the amount of installed RAM DIMMS but need a way to total them up and produce a varible I can use in the pre section of the install.
Note: on some servers there could be from 1 DIMM up to 16 DIMMS installed so the script needs to be able to handle this. I also can not use bc as it does not exist during the install stage. I am guessing I need a while loop to do this and use expr but do not know where to start for this logic.
I just noticed that one of my blade server has such a abnormal running status.Historically, the three values in file-nr denoted the number of allocated file handles, the number of allocated but unused file handles, and the maximum number of file handles.But it's really strange to see low value for the first column.
Just wondering now I see computers everywhere with 6G of ram or more if the new 32b kernels are going to be setup for large amounts of ram or will they still need to be recompiled? Also will there be better ssd support for things like "trim"? I know you edit the fstab to get the Linux version of trim but I am just wondering if stuff like that will just be automatic or if I will continue to need to tweak. I like to tweak but sometimes its the lack of need for tweaking that makes me like slack.
I used 9.04 for months and it work fine before restarting my PC. After I restarted my PC, the memory consumption takes up to 4.2 GB after login. However, I cannot find any process that consume such large number of memory.
I want to transfer an arbitrarily large file (say >20GB) between 2 servers. I have several considerations:
Must use port 22 (ssh) because of firewall restrictions Cannot tax the CPU (production server) Memory efficiency Would prefer a checksum check but that could be done manually Time is not of the essence
Server A and Server B are on the same private network (sharing a switch) and data security is not a concern, Server A and Server B are not on the same network and transfer will be via the public internet so data security is a concern, My first thought was using nice on an scp command with a non-CPU-intensive cypher (blowfish?). But I thought I'll refer to the SU community for recommendations.
I have a few multi-user servers in an academic laboratory. I am having a problem with some users maxing out the available RAM, causing such sever slowdowns the machine essentially crashes. My servers are Dell Power Edge's running Ubuntu 8.10 Server Edition (Not my choice). I would like to set a maximum limit on the amount of ram a user can utilize. This morning I experimented with setting limits via /etc/security/limits.conf and using ulimit. Neither of them prevented my test program, a simple infinite loop of mallocs, from crashing the server.
I am connected to a network via SHH. Now, i know how i can see which users are logged on as well, but how can i see how many times each user has connected? this refering to users that have logged on at least once
Is there a system admin tool for RHEL on montioring and logging system memory used and released that can dump to a log file?I'm having an issue with memory not being released when an application is closed. I need to have a tool monitor and log so I can troubleshoot to verify that it's an application not the OS.
Is there a file system that both Mac OSX 10.5 and linux can read/write for large files (like 4gb files)? My desktop is Ubuntu and I run most from there, but I want to back up my MacBook and linux box on the same external hard drive. Seems there are some (paid) apps for Mac that will mount NTFS but I'm wondering if there is just a shared files ystem that will work for both.
I have a 58 mg PDF that needs to be sent to a person with Windows. I do not know what special software they have. I only know that I nust send them the file in a way that they can use it. I tried yousendit.com and it tells me I need to install their "yousendit express package". But apperantly it will not run on Linux.
I am having a bit of a problem with my Ubuntu Server 10.04 install. I think it might be a kernel problem. Basically, what happens is when I copy a large file (a 160GB disk image) to my drive (>60GB) the system consistently crashes after about 60GB of the file is transferred. It doesn't matter if I am sending the file using cifs, or over SSH. Checking syslog (paste dump here), it seems these flush errors always appear shortly before the crash occurs. The destination filesystem is a hardware RAID 10 array with 2TB of space. It is formatted as EXT4.