Ubuntu :: Understand A Large Amount Of Allocated Memory That Seems Not To Be Accounted For On System?
Mar 24, 2010
I am trying to understand a large amount of allocated memory that seems not to be accounted for on my system.I'll say up front that I am discussing memory usage without cache and buffers, 'cause I know that misunderstanding comes up a lot.I am in a KDE 4.3 desktop (Kubuntu 9.10), using a number of java apps like Eclipse that tend to eat up a lot of memory.after a few days, even if I quit most apps, 1 gb of ram remains allocated (out of 2 gb).this appeared excessive, and I took the time to add up all values of the RES column in htop (for all users).the result was about 1/2 gb.am I trying to match the wrong values?or could some memory be allocated and not show up in the process list?this is the output of free
Code:
total used free shared buffers cached
Mem: 2055456 1940264 115192 0 123864 702900
[code]...
View 6 Replies
ADVERTISEMENT
Aug 25, 2010
Is there an easy way in C to get the total amount of allocated SHM memory on a machine?
View 7 Replies
View Related
Nov 8, 2010
I need to allocate a % of the total system memory for a buffer but what is the best method to determine how much memory is in the system? So far the only way I have found is to get the pages of memory:
Code:
long sysconf(_SC_PHYS_PAGES)
Is that the only option?
View 2 Replies
View Related
Jul 15, 2010
I want to calculate the total amount of memory used by the system, Someone please confirm if the below formula is correct.Quote:
((memUsed-mebBuffers-memCached)*100)/memUsed => totalMemoryUsed%
Code:
free -o
total used free shared buffers cached
Mem: 8167004 8115292 51712 0 87680 1444676
Swap: 8388584 3252996 5135588
((8115292-87680-1444676)*100/8115292) => 81%
As per the above calculation 81% of memory is used.Is this correct? and if so Am I running out of memory?what is the limit in % that I should maintain for a better performance?
View 3 Replies
View Related
Aug 18, 2010
I am still confused about the memory issues. i have allocated JAVA_OPTS='-Djava.awt.headless=true -Xms512M -Xmx512M -XX:MaxPermSize=512M' think that it means that i have allocate java to use only 512mb of my memory. but when i run top here is my output.
PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
15932 root 23 0 1495m 1.2g 3732 S 12.0 36.0 121:52.98 java
5070 postgres 15 0 55392 39m 35m R 11.6 1.2 0:51.61 postmaster
[code]....
View 6 Replies
View Related
Jun 1, 2010
I want to know if i can increase the memory space allocated for a process manually while the process is running ,,,, and if it is possible how i can do this .
View 2 Replies
View Related
May 20, 2010
Well, when I copy large amount of data the other applications than Nautilus freezes until the copy is done...
So, what can I do? Because when backuping some data this is really annoying =/
View 6 Replies
View Related
Mar 5, 2011
In gparted I have the following stats for my /home drive
size: 824 gb
used 75.51 gb
unused 748.59 gb
Now when I view this in nautilus it shows something else: remaining free space as 709 gb. My question is what happened to the 40gbs? the 75.51gb are my files, but where did the 40gbs go to? Because 709 (total remaining) + 75 (my files) + 40 (mysteriously lost gbs) = 824gb. When I first made the partiton, it was a 824gb partition and ubuntu had automatically at that point reserved about 40gb for something. Does anyone know why Ubuntu reserved this space?
View 4 Replies
View Related
Aug 8, 2011
Let's analyse this While registering a Char deriver to kernel think that we had allocated memory using any of malloc/kmalloc/vmalloc.
If we forgot to Free that allocated memory while unregistering, will that memory be allocated aside or kernel clears it...?
View 2 Replies
View Related
Dec 11, 2010
I use Audacious to play my music... But every time I want to add my music it crashes... even when I do it in batches. Is there somewhere I can find logging to see what the problem is?
View 7 Replies
View Related
May 14, 2009
On a new box with Debian Lenny, I have 2x2 gig of ddr2 but the karamba applet show only 3291mb. And dmesg show this:
Code:
dmesg | grep Memory
[0.004000] Memory: 3362976k/4194304k available (1769k kernel code, 43092k reserved, 752k data, 244k init, 2489792k highmem)
View 2 Replies
View Related
Jun 4, 2010
as you should see, top is indicating 3.544.864kb (3.5Gb) of memory used while gnome system monitor only 609Mb. What's wrong here? (I am pretty sure Gnome SM is right. Top is updating every sec.)
View 1 Replies
View Related
Mar 3, 2009
I would like to know if there is a linux command to verify the amount of memory used by a program. The programs I am using were compiled with gfortran.
View 2 Replies
View Related
May 17, 2011
My problem is I installed Zone Minder for camera security and I'm testing it on my laptop with the built in webcam and everything seems to work perfectly except when I try to view the live feed from the camera, it's just a black box. No video.
I checked this website and it's exactly the problem I'm having with a fix for it but his fix doesn't work. He says to type:
Code:
user@ubuntu:~$ sudo echo "256000000" > /proc/sys/kernel/shmmax
user@ubuntu:~$ sudo service apache2 restart
user@ubuntu:~$ sudo service zoneminder restart
But when I do the first line, I get this:
Code:
deemar@Clementine:~$ sudo echo "256000000" > /proc/sys/kernel/shmmax
bash: /proc/sys/kernel/shmmax: Permission denied
View 2 Replies
View Related
May 9, 2011
When I run ardour sound editing I get this message , but it starts ok Your system has a limit for maximum amount of locked memory! This might cause Ardour to run out of memory before your system runs out of memory. You can view the memory limit with 'ulimit -l', and it is normally controlled by /etc/security/limits.conf
bash-4.1$ ulimit -l
64
my limits.conf is like this
audio - rtprio 99
@audio - memlock 250000
[Code]...
View 2 Replies
View Related
Apr 4, 2011
I am trying to write a script to calculate the total amount of installed memory to use during an anaconda kickscript, so the swap file is created at 2 x the installed memory. I so far have the amount of installed RAM DIMMS but need a way to total them up and produce a varible I can use in the pre section of the install.
Code:
dmidecode -t 17 | grep Size | awk ' { print $2 }'
Output:
2048
2048
Note: on some servers there could be from 1 DIMM up to 16 DIMMS installed so the script needs to be able to handle this. I also can not use bc as it does not exist during the install stage. I am guessing I need a while loop to do this and use expr but do not know where to start for this logic.
View 3 Replies
View Related
Mar 27, 2010
Was wondering if anyone can explain briefly the relationship of "cache" and free memory in the "free" memory command.
View 2 Replies
View Related
Apr 22, 2010
I just noticed that one of my blade server has such a abnormal running status.Historically, the three values in file-nr denoted the number of allocated file handles, the number of allocated but unused file handles, and the maximum number of file handles.But it's really strange to see low value for the first column.
View 2 Replies
View Related
Mar 29, 2011
(inside R) I am using system.time function to measure the execution time of two functions
> system.time(lapply(seq(1:1000000),returni))
user system elapsed
0.982 0.100 1.057
[code]....
View 3 Replies
View Related
Jan 4, 2011
Just wondering now I see computers everywhere with 6G of ram or more if the new 32b kernels are going to be setup for large amounts of ram or will they still need to be recompiled? Also will there be better ssd support for things like "trim"? I know you edit the fstab to get the Linux version of trim but I am just wondering if stuff like that will just be automatic or if I will continue to need to tweak. I like to tweak but sometimes its the lack of need for tweaking that makes me like slack.
View 14 Replies
View Related
Feb 23, 2010
I used 9.04 for months and it work fine before restarting my PC. After I restarted my PC, the memory consumption takes up to 4.2 GB after login. However, I cannot find any process that consume such large number of memory.
[code]....
View 3 Replies
View Related
Oct 5, 2010
I want to transfer an arbitrarily large file (say >20GB) between 2 servers. I have several considerations:
Must use port 22 (ssh) because of firewall restrictions
Cannot tax the CPU (production server)
Memory efficiency
Would prefer a checksum check but that could be done manually
Time is not of the essence
Server A and Server B are on the same private network (sharing a switch) and data security is not a concern, Server A and Server B are not on the same network and transfer will be via the public internet so data security is a concern, My first thought was using nice on an scp command with a non-CPU-intensive cypher (blowfish?). But I thought I'll refer to the SU community for recommendations.
View 1 Replies
View Related
Jul 24, 2009
I have a few multi-user servers in an academic laboratory. I am having a problem with some users maxing out the available RAM, causing such sever slowdowns the machine essentially crashes. My servers are Dell Power Edge's running Ubuntu 8.10 Server Edition (Not my choice). I would like to set a maximum limit on the amount of ram a user can utilize. This morning I experimented with setting limits via /etc/security/limits.conf and using ulimit. Neither of them prevented my test program, a simple infinite loop of mallocs, from crashing the server.
View 7 Replies
View Related
Feb 23, 2011
Fresh install of Debian 6.0 (squeeze) and the memory shows 3.8 GB. I have 6 sticks of 2 gb on the board.
Pertinent dmesg lines: AMI BIOS detected: BIOS may corrupt low RAM, working around it.
3083MB HIGHMEM available.
883MB LOWMEM available.
Bios tests 12 gb of ram fine.
View 4 Replies
View Related
Apr 9, 2011
How to get output of text file containing account number, debit amount, credit amount,date using shell script?
View 1 Replies
View Related
Mar 11, 2010
I am connected to a network via SHH. Now, i know how i can see which users are logged on as well, but how can i see how many times each user has connected? this refering to users that have logged on at least once
View 7 Replies
View Related
Apr 20, 2011
Is there a system admin tool for RHEL on montioring and logging system memory used and released that can dump to a log file?I'm having an issue with memory not being released when an application is closed. I need to have a tool monitor and log so I can troubleshoot to verify that it's an application not the OS.
View 2 Replies
View Related
Sep 19, 2010
Is there a file system that both Mac OSX 10.5 and linux can read/write for large files (like 4gb files)? My desktop is Ubuntu and I run most from there, but I want to back up my MacBook and linux box on the same external hard drive. Seems there are some (paid) apps for Mac that will mount NTFS but I'm wondering if there is just a shared files ystem that will work for both.
View 9 Replies
View Related
Jan 18, 2011
I have a 58 mg PDF that needs to be sent to a person with Windows. I do not know what special software they have. I only know that I nust send them the file in a way that they can use it. I tried yousendit.com and it tells me I need to install their "yousendit express package". But apperantly it will not run on Linux.
View 9 Replies
View Related
Jun 15, 2010
I am having a bit of a problem with my Ubuntu Server 10.04 install. I think it might be a kernel problem. Basically, what happens is when I copy a large file (a 160GB disk image) to my drive (>60GB) the system consistently crashes after about 60GB of the file is transferred. It doesn't matter if I am sending the file using cifs, or over SSH. Checking syslog (paste dump here), it seems these flush errors always appear shortly before the crash occurs. The destination filesystem is a hardware RAID 10 array with 2TB of space. It is formatted as EXT4.
View 7 Replies
View Related