Ubuntu :: Takes Too Much Memory Even If All Processes Killed?

Jan 9, 2010

Here's what i have with Ubuntu 9.10:

Code:
ruslan@hplaptop:~$ ps -Ao%mem,pid,cmd
%MEM PID CMD

[code]...

View 14 Replies


ADVERTISEMENT

General :: Top Used Memory Versus Ps Processes Memory

Jan 17, 2010

I found from command 'top' that 8GB memory are used. However, using command 'ps' with some options to grep the running processes and then summing up the memory used by the running processes are less than 2 GB. Where has the used memory gone ?

View 7 Replies View Related

Ubuntu :: Gvfsd Takes Up Too Much Memory?

May 13, 2010

On my Ubuntu 8.10, gvfsd takes up 128.6MB memory. What is the process doing actually? Can I reduce the memory it takes?

View 2 Replies View Related

Ubuntu Installation :: Firefox 5 Takes All Available Memory

Jul 22, 2011

a few months ago i did a fresh install of ubuntu 11.04 64-bits. it works perfectly. i'm using firefox 5.0, and in about it says mozilla firefox for ubuntu canonical 1.0.

every now and then, firefox behaves like a runaway train and starts using up all available memory, and using 100% processor time. this has been going on for a week or so. the only remedy is to kill firefox and restart it.

i'm not sure what causes this problem, but i am starting to suspect that it may be caused by flash, java, or some other plug-in or add-on that i have installed.

my add-ons are:
- adblock plus 1.3.9
- adblock plus pop-up addon 0.2.7
- delicious bookmarks 2.3.1
- delicious extension 2.3.1
- flash-aid 2.1.1
- global menu bar integration 1.0.6
- memory restart 1.4
- stumbleupon 3.91
- ubuntu firefox modifications 0.9.1

with the exception of memory restart 1.4 all these add-ons have been installed for a long time, without ever causing any problems. the problems started before installing memory restart.

my java version is icedtea, installed from the repository. i'm frequently using sweet home 3d (a java program) that runs without any problem. it doesn't make a difference whether i have sh3d running or not, the ff problem occurs regardless.

i also have installed ubuntu restricted extras version 43. videos videos don't cause problems. vimeo videos cause problems on occasion when switching to high definition and/or to full screen, but that is a known vimeo problem.

the problems certainly start when i don't use my computer for a longer period of time. i have disabled the screen saver/energy saver, but that doesn't seem to solve the problem. today i had the problem shortly after visiting a website, reason why i'm starting to suspect that the problem could be linked to some plug-in that website uses.

View 9 Replies View Related

Server :: SNMP NOTIFICATION-TYPE - Memory: Killed Process 8005 (snmpd)

May 5, 2011

My Objective is to add Private MIB and compile that file and perform snmpget and snmpset operation. My MIb file contain this macro

[Code]...

Next I try to cmpile using mib2c highCpuUsage. it's give template code for that. I read all the tutorial and copy in /mibgroup and restart snmpd service and try snmpwalk ,snmpset but it�s not work and in var/log/messages its give Out of Memory: Killed process 8005 (snmpd)

View 2 Replies View Related

OpenSUSE :: Kde 4.4 Takes Almost Entire Memory?

Oct 20, 2010

Code:
total used free shared buffers cached
Mem: 997 950 47 0 21 805

How can i reduce used memory?

openSUSE 11.3 (i586)
VERSION = 11.3

View 5 Replies View Related

Ubuntu Servers :: Rsyslog Takes 100% Of Memory On Boot?

Jun 15, 2010

I'm trying to run a logging server with encryption but rsyslog takes 100% of the memory on boot. This only happens when these two sets of lines are both in the rsyslog.conf

Code:
$ModLoad imtcp
$InputTCPServerRun 10514
and
Code:
$DefaultNetstreamDriver gtls

[Code]...

View 2 Replies View Related

General :: Getting The Multi-processes Memory Used?

Sep 2, 2010

The following code is for monitoring the memory used by apache processes. But I got a problem that the data I got by this script is much larger than the physical memory. It was said that there are some libraries used simultaneously by many processes, so the data I got has some double counted part. Because apache has many httpd processes.

Anyone have an idea of getting the multi-processes memory used?

Code:
#!/bin/sh
#G.sh 20100813
USAGE="Usage: $0 processName"

[Code].....

View 14 Replies View Related

Software :: All Memory Being Consumed By Only A Few Processes?

Jan 25, 2010

I recently built a new server, running 64 bit Slackware 13.0 with the following specs:

MSI 785GTM-E45
AMD Phenom II X2 550
2GB DDR2
Onboard video from AMD 785G chipset
2x 80GB IDE system drives using software RAID with 2GB swap partition

I only include these because I'm not convinced my problem is not hardware related at some level. Basically, when I first start up the system, the memory usage is anywhere from 60 to 200MB. Then it starts to gradually climb until there is only 12-15MB free. This can take anywhere from a few hours to a few days.

The only thing I really use this for is to serve Samba shares and the occasional SSH login. I've set up Samba shares to be accessed by my Windows machines as well as a Mac. Initially I just explored the network if I wanted to see the share from these machines, but that was too slow and unreliable (the server would not always show up in Windows), so now I automatically mount the share as a network drive at startup (from Windows). So I don't know if this would have anything to do with the steadily increasing memory usage. These systems are not on/connected all the time, but the memory usage seems to rise anyway.

When I run top, it reports that nearly all of the physical memory has been consumed (after a while of uptime), but none of the swap space has even been touched. This is a typical output of the first several lines, sorted by swap size:

Code:

top - 07:45:48 up 2 days, 12:33, 2 users, load average: 0.00, 0.00, 0.00
Tasks: 105 total, 1 running, 104 sleeping, 0 stopped, 0 zombie
Cpu(s): 0.0%us, 0.2%sy, 0.0%ni, 99.8%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st

[code]....

I can't be sure, but I believe the message just repeated itself before this part. MS-7549 is one of the components of the mainboard (MSI), maybe the northbridge or something (it seems to be associated with more than one model), so I thought for sure this meant a hardware problem, maybe with the storage controller or memory. But Memtest86 passes for more than a day without any errors, and I loaded Windows on one of the disks and ran it as hard as I could with prime95 and nothing so much as flinched. So I've got to think this is some kind of issue with linux and/or how it is handling my hardware configuration.

View 5 Replies View Related

Programming :: Memory Optimization For Child Processes?

Sep 6, 2010

I work on Linux for ARM processor for cable modem. There is a tool that I have written (as the job demands) that sends/storms customized UDP packets using raw sockets. I form the packet from scratch so that we have the flexibility to play with different options. This tool is mainly for stress testing routers.

The details are here.

I actually have multiple interfaces created. Each interface will obtain IP addresses using DHCP. This is done in order to make the modem behave as virtual customer premises equipment (vcpe).

When the system comes up, I start those processes that are asked to. Every process that I start will continuously send packets. So process 0 will send packets using interface 0 and so on. Each of these processes that send packets would allow configuration (change in UDP parameters and other options at run time). Thats the reason I decide to have separate processes.

I start these processes using fork and excec from the provisioning processes of the modem.

The problem now is that each process takes up a lot of memory. Starting just 3 such processes, causes the system to crash and reboot.

I have tried the following:- 1-I have always assumed that pushing more code to the Shared Libraries will help. So when I tried moving many functions into shared library and keeping minimum code in the processes, it made no difference to my surprise.

2-I also removed all arrays and made them use the heap. However it made no difference. This maybe because the processes runs continuously and it makes no difference if it is stack or heap?

3-I suspect the process from I where I call the fork is huge and that is the reason for the processes that I make result being huge. I am not sure how else I could go about. say process A is huge -> I start process B by forking and excec. B inherits A's memory area. So now I do this -> A starts C which inturn starts B will also not help as C still inherits A?. I used vfork as an alternative which did not help either. I do wonder why.

reduce the memory used by each independent child processes.

View 1 Replies View Related

General :: Get The Physical Memory Used By A Couple Of Processes?

Sep 3, 2010

I am doing a test to get the memory used by apache`s apache2 processes. I used a script to get VmSize and VmRss from /proc/pid/status, and loop through that to get the sum of VmSize and VmRss of all the apache2 processes.

I found the VmSize (about 4GB) and VmRss (about 3.4GB) are much larger than the physical memory (1GB) when apache server was saturated. It was said because of the multi-counted libiraries size used by many processes simultaneously. Then , how to get the physical memory used by apache2 processes? Or how to get a more reasonable memory data?

View 7 Replies View Related

General :: Memory Optimization For Child Processes?

Sep 6, 2010

I work on Linux for ARM processor for cable modem. There is a tool that I have written (as the job demands) that sends/storms customized UDP packets using raw sockets. I form the packet from scratch so that we have the flexibility to play with different options. This tool is mainly for stress testing routers.I actually have multiple interfaces created. Each interface will obtain IP addresses using DHCP. This is done in order to make the modem behave as virtual customer premises equipment (vcpe).

When the system comes up, I start those processes that are asked to. Every process that I start will continuously send packets. So process 0 will send packets using interface 0 and so on. Each of these processes that send packets would allow configuration (change in UDP parameters and other options at run time). Thats the reason I decide to have separate processes. I start these processes using fork and excec from the provisioning processes of the modem. The problem now is that each process takes up a lot of memory. Starting just 3 such processes, causes the system to crash and reboot.

I have tried the following:-

1-I have always assumed that pushing more code to the Shared Libraries will. when I tried moving many functions into shared library and keeping minimum code in the processes, it made no difference to my surprise.

2-I also removed all arrays and made them use the heap. However it made no difference. This maybe because the processes runs continuously and it makes no difference if it is stack or heap?

3-I suspect the process from I where I call the fork is huge and that is the reason for the processes that I make result being huge. I am not sure how else I could go about. say process A is huge -> I start process B by forking and excec. B inherits A's memory area. So now I do this -> A starts C which inturn starts B will also not help as C still inherits A?. I used vfork as an alternative which did not help either. I do wonder why.

View 1 Replies View Related

Server :: Can't Figure Out Which Processes Are Eating Memory

Nov 6, 2010

I'm running several SHOUTcast server instances and a WowzaMediaServer instance on a CentOS machine. I'm experiencing a memory leak problem, but I can't figure out which processes are eating memory.

TOP command reports as follows:

[Code]...

Something misterious to me (I'm still a Linux newbie) is that TOP reports a total of 7.5GB used ram but very small percentage for single process (0-1%). Memory consumption starts at 1GB/8GB after reboot and in three days running gradually increases up to 8GB. I'm practising with Linux, but I still miss a lot to understand what's happening on my system. For instance, are there linux kernel logs saved somewhere that I can look at?

View 4 Replies View Related

Programming :: Synchronization Of Two Processes With Share Memory?

Feb 27, 2011

I have two processes that share a piece of memory, and i want to use the shared memory to send data from one process to the other. it's like a simple consumer-producer problem. when the producer fills the shared memory, it waits until the consumer can consume some data in the memory; the consumer needs to wait if there is no data in the memory. The thing gets complicated when both threads are allowed to sleep and wait for the other to wake it up.

i wanted to use condition variable of pthread for synchronization, but it doesn't work in multiple processes. i tried semaphore, but it's quite complicated and i still cannot make it right. I believe it's a common problem and someone should have written similar code before, or maybe the code is even wrapped in a library, but when I search for it on Internet, I only found information about how to share memory between processes. Does anyone know where I can find this kind of code or library?

View 2 Replies View Related

Server :: Controlling Memory/number Of Processes For LDAP Users?

Jun 29, 2009

I know of /etc/security/limits.conf and that can be used to limit all sorts of good things, but I haven't found anything that talks about using this when the users come from LDAP. Would I be able to do something like

@"Domain Users" soft nproc 25
@"Domain Users" hard nproc 40

where Domain Users is the group all users belong to in our system.

View 3 Replies View Related

General :: Child Processes Of A Parent Process Occupying Socket Memory?

Jul 22, 2010

I am facing an issue where the process starts hanging. When I closely look at the logs I come to know that some of the child processes that are forked by the parent process are not finished.

1) Is it possible that the child processes that are not finished occupy the socket memory of the parent process and ultimately a point is reached where no socket memory is available to fork new child processes.

2) What is the standard limit of socket memory in linux?

3) What is the fate of such child processes (as I have mentioned above)?

4) How to debug such cases so that the exact problematic area is identified?

View 2 Replies View Related

General :: Tools To View/edit User Space Memory Of Running Processes?

Nov 2, 2010

Are there any tools to view/edit user space memory of running processes on Linux?

It would be a great learning tool.

View 1 Replies View Related

General :: Wrote A Program That Multiplies 2 Matrices Using Multi-threads And Another One Using Multiple Processes And Shared Memory?

Sep 22, 2010

I wrote a program that multiplies 2 matrices using multi-threads and another one using multiple processes and shared memory. Both in C.I need to find the total memory usage of these programs. I know of the top command, but when my matrices are relatively small they don't even show up on top because they complete so fast, how can I find the memory usage for these instances?Also, how can I find the total turnaround time of my programs

View 1 Replies View Related

CentOS 5 :: X/Gnome Takes Long Time To Load And Sudo Takes Long Time To Execute

Oct 13, 2009

I am running Centos 5.3. I ran no updates, performed no installs, nor changed any configuration immediately prior to this issue. My problem is this: when I run the command startx (default runlevel 3), it is a long time (5-10 minutes) before Gnome startx, and once it does start applications will not run. Also, when I try to use sudo (from any environment, even ssh), it is a long time (5-10) before the command is executed.

I cannot say for sure, but it seems like this is an intermittent problem. Sometimes X takes a long time to start, but once it starts it will launch programs. Sometimes X takes a long time to launch, but once it starts it will only launch certain programs. Though presently X always takes a long time to start, and I cannot successfully launch any programs.

A while back a had a similar problem to this (x taking long time to start, sudo taking long time to execute) and it ended up being a DNS problem. Unfortunately, I cannot remember exactly what it was and I stupidly did not document it. Maybe this is also DNS related, I don't know.

I don't know what log files to look at for problems with X, Gnome, and sudo taking a long time to start.

View 19 Replies View Related

General :: Ubuntu - Memory Full, Process Memory Usage Doesn't Add Up In Top?

Mar 18, 2011

I have a computer with 16GB of ram. At the moment, top shows all the RAM is taken, (NOT by cache), but the RAM used by the various processes is very far from 16GB.I have seen this problem several times, but I don't understand what is happening.My only remedy so far has been to reboot the machine.

View 1 Replies View Related

Ubuntu :: Memory Error When Installing Python - Cannot Allocate Memory

Apr 19, 2010

I get this error when I run "sudo apt-get install python-software-properties"

Preconfiguring packages ...
dpkg: unrecoverable fatal error, aborting:
fork failed: Cannot allocate memory
E: Sub-process /usr/bin/dpkg returned an error code (2)

I'm trying to install deluge via ssh and my vps has 512mb ram and is only using 11% of it prior to running the code.

View 2 Replies View Related

Ubuntu :: Memory Leak / Some Process Eating All Memory

Jan 29, 2011

I have had a fresh install of Ubuntu 9.10 and installed some software after that.Since third some, some process is eating half of my memory.I have checked processes running in system manager but everything is normal.Maximum is consumed by compiz which is about 26 mb, seems very normal.I did restarted my computer several times, and in the start for 5 mins, its fine after that again my cpu fans runs at very fast speed and my one cpu is used up 95 % (I have dual core).Please help me out, this invisible thing is driving me crazy.I am attaching my htop screen shot (sorted by cpu %), now the cpu is not used by completely but fan is still struggling hard and fast.

View 9 Replies View Related

General :: Application Virtual Address Space Memory Allocation - Memory Does Not Get Free

Apr 20, 2010

I am using malloc and frees a lot in my program. It shows its allocated but when i remove it doesnt show as the memory is removed(I am using the top command to view VIRT memory usage). If this continously grows what would happen to my program (Will it go out of memory?)

View 4 Replies View Related

Fedora Hardware :: Checking Computer Memory - Specification Shows 2 GB Of Memory Instead Of GB

Feb 7, 2011

Fedora 14 xfce
HP Mini 210

I am looking to buy some memory for my netbook. Currently I have 1 GB of DDR3 memory. However, the specification says that 2 GB of memory is the max. However, when I do the following it says that 4GB is the max:

[Code].....

View 3 Replies View Related

Server :: Clear Cache Memory ( RHEL 5.1 ) As It Consumes Almost 100% Physical Memory?

Jan 11, 2010

let me know how to clear cache memory ( RHEL 5.1 ) as it consumes almost 100% physical memory.

View 3 Replies View Related

Software :: Free Database That Has Low Memory Usage And Innodb And Memory Like Engin

May 3, 2011

I am looking for free database that has low memory usage and innodb and memory like engins that has C API and support trigger and client/server support for using in embedded linux systems.

View 8 Replies View Related

Programming :: Memory Leaks - Glibc Detected - ./SuffixTree: Malloc() - Memory Corruption - 0x00

Mar 21, 2011

I am new to C and linux. My code below does arbitary writes but I cant figure out where or how it does it.

I am calling the insertNode() function with seq = 'MISSISSPPI$' and alphabets = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ$'

Code:

Weird behaviour I should mention is that when I check for NULL pointer in node->child[index], the unassigned values are not null anymore, they point to arbitary memory.

View 12 Replies View Related

OpenSUSE Install :: Virtual Memory / Resident Memory Of Server?

Feb 3, 2010

we found that if we use 'top' to show the memory usage of a server (SuSe Linux 10), we can get virtual memory usage as well as 'Resident memory' usage. For virtual mem or a particular process, it is around 1.1GB, which is large but for resident memory, it only consumes 300MB. Are there anyone who knows what the differences are? I would also like to know whether the difference (1.1GB - 300MB) = 800MB are actually available for use by other applications in the system.

View 1 Replies View Related

Programming :: Write A Script That Can Show Me Total Memory Vs Used Memory?

Apr 13, 2010

How do I write a script for my Linux that can show me total memory vs used memory and have it email me results if it's over 70 percent?

View 2 Replies View Related

Hardware :: Why Dmesg / Grep Memory Gives Two Different Physical Memory Amounts

Oct 20, 2010

I am monitoring physical memory in a server I administer, and my hardware provider told me they had increased physical memory size to 4Gb... However, using several tools (free -m; top; dmesg | grep Memory; grep MemTotal /proc/meminfo I discovered that I actually have 3Gb, not 4... But, my doubt comes from the fact that dmesg | grem Memory tells me I have 3103396k/4194304k available The first number is effectively 3Gb, but the second one, is 4! so, why I am looking at this two different numbers?

View 1 Replies View Related







Copyrights 2005-15 www.BigResource.com, All rights reserved