I've installed my debian sid about one month ago (first xfce, next gnome) but noticed that it's kind of really slow. The upgrades take ages, launching (and using) firefox takes so much time,... In comparaison to my ubuntu, archlinux (on the same computer) or previous installation of debian there is clearly a problem somewhere.Today I tried to do a "top" sorted by mem usage : 3.5% xulrunner-stub, 2.1% dropbox, 1.4% aptitude (doing upgrade), 1.4% clementine,... nothing terriblebut still I've 2.7Gb or RAM used (more than 50%)
My problem seems to be very simple, it's high memory usage. I occasionally will use movie player to watch a few shows and I use firefox as well. My memory usage starts out real small about 500 mb but after using firefox lightly and movie player it jumps to almost 2 gigs and this is after they've been closed what gives? I've attached an image so you can see what I'm talking about.
I upgraded from Fedora 13 to 14 over the network. Everything seems to have worked. The one problem after my install is that I have noticed that setroubleshootd consumes alot of memory.
It doesn't take long for setroubleshootd to jump in memory usage. I can kill the process but it will start up again. I have tried disabling the service but it doesn't show up in /etc/init.d. # service setroubledshootd stop setroubledshootd: unrecognized service So I am not sure what I can do to resolve the issue with setroubleshootd besides killing it off every 15 minutes.
In our database, when checking the memory used using top command, we are always seeing 32 GB RAM utilized. We have set the sga_max_size to 8gb and PGA to 3 gb. We have tried shutting down oracle db and then the memory went down to 24GB when checked using top command. After cold reboot of the DB server, it gone down to 1.5 GB.
But once the users are started using after end of day, the memory again gone back to 32 GB.
i am having a problem that i would call a bit "important" with my server. so, from last 3 weeks the used space of my hard disk (RAID I) started growing up. i have 2 x 1 tb HDD working on RAID I and i did not install anything those weeks. the space just started changing from 90 GB till 580 GB. now the situation is stable there but i think it's not normal.
the bandwidth usage is low (like 120 gb in 2 months) and i am running 6 counter strike gameservers, a forum, a very little website and some local stuffs... a friend of mine told me that my server could have been hacked but i am afraid it did... some useful informations: when i reboot the server the used space goes down again to ~100 GB and then it starts going up again. i cant really find where all those files are located:
Basically I have a machine with 16GB of RAM and have just discovered that using all of it can crash the whole system over one process. How could I run a process on the system in such a way that if more than 90% of system memory is used, the process immediately crashes?
I recently upgraded to Ubuntu 10.10 and I am experiencing an ultra high memory usage of the gnome-settings-daemon of 2GB after suspend! Killing and restarting the daemon solves the issue. Anybody else with this behavior?
I was browsing my folder with lots of images, after finished i close nautilus and i notice that my computer became slow, so i'll check it with system monitor and had found that nautilus are using almost 100mb of ram (opening 4 tabs). I'm not sure if this was normal or not because i try to reopen the same folder with pcmanfm and it only consumes less than 20mb of ram (opening 4 tabs). here's the screenshot from system monitor .
I have two pc's that I've recently converted to Lenny boxes, and I've noticed that whenever there's a lot of disk IO, one of my cores (both pc's have dual core cpu's) jump to 100%. Is this normal? They're both amd64, and I think both of them should be using the AHCI SATA driver.
If you need any more information about my setup, let me know.
I recently switched from jessie to stretch on both i encountered a jerky ui in gnome 3. after a longer research i noticed xorg consumes over 80% of my first cpu-core while moving a window. for testing purposes i switched back to the nouveau driver and there is no high cpu usage while moving a window, also the ui is smooth.
i tried 3 different nvidia driver: the one from the debian repository, the current stable and beta driver from the nvidia site. i noticed on all 3 the same problem.
I use a Debian Squeeze system running off a flash drive, i.e. based on a custom Live image running in persistent mode. It runs great and I am grateful for the existence of Debian . However, I have a question. A lot of the machines I use this pen drive on are quite old, often with 512 MB RAM and old processors. I specifically built my system using XFCE and lightweight apps off an initial live image using the standard-x11 package list (basically just Xorg with drivers and the base system). At first things ran very well, blazing fast even on the oldest systems and could comfortably run Firefox along with LibreOffice side by side (I need LO as all of my colleagues use Word docs, often with track changes, which Abiword can't handle properly). However, over time, I've found that memory usage has risen, tot he point where Firefox is now automatically killed on the older systems every time I start LibreOffice.how does one figure out why memory usage is going up? I've checked for inessential services and turned them off with "insserv -r". I've used only lightweight apps, as mentioned before. Are there other general tips on reducing memory usage?
I have a home server set up with samba, proftp, and the torrentflux webUI via apache and mysql. To the best of my ability, I have stripped the system of any unnecessary software. After a fresh reboot the system uses around 250 mb of memory, which to me still seems like a lot, and after it has been on for a few hours with torrents downloading, the memory usage will steadily climb to and stay at 2 gb! Using top shows that the processes using the most memory are python, mysqld, and apache2. What the heck is going on?
Is that possible that SHM shared memory is counted as cache memory on Linux with kernel 2.6.18?If find it really odd since this memory is not file backed, but I have a piece of code that loads data using shm_open+mmap, and it generates an amount of cache memory in /proc/meminfo that corresponds exactly to the amount of shared memory (I load that data from a file but I am using posix_fadvise(fd,0,0,POSIX_FADV_DONTNEED) to ensure this file is not cached and I made sure that it is working as expected). As far as I know SHM memory was not tagged as cache memory with kernel 2.6.9.If it is the case it is really unfortunate since normally cache memory can be considered to be part of the "available" memory since it can be flushed promptly but this is clearly not the case with SHM memory... Is there an easy way to get the total amount of used SHM memory on a system?
I was trying to get the status of memory usage and disk usage using sigar in windows and ubuntu. done this in windows by just copying the sigar library into jdk library. But i was unable to do so in ubuntu. I've copied the library to java-6-sun library but still can't run the program.
I have a java program that runs on Debian as a background processor. Yesterday the Java program stopped running. I looked at the memory usage, the system only had 5MB memory left, so my guess is that the java program ran out of memory to use.
However, after we restarted the java program, we could see that the free memory count started to go up. It kept going up from 5MB to over 400MB. The increase of memory happened slowly, when I measured it, I could see that with each minute passing by, there were a bit more memory added into the free memory pool, and meanwhile, the java background process was running.
I wonder why this would ever happen. It's as if our java program first brought the machine done because it consumed all the memories, then after restart, it starts to give back memories.
I have a computer with 16GB of ram. At the moment, top shows all the RAM is taken, (NOT by cache), but the RAM used by the various processes is very far from 16GB.I have seen this problem several times, but I don't understand what is happening.My only remedy so far has been to reboot the machine.
I've come across a really strange issue with one of my RHEL servers. The "free" command shows that 7019 MB of memory are actually in use by my system, but when summing up the actual usage (or even virtual usage like the example below) it doesn't add up - the sum is far less than what is reported by "free":
I just upgraded KDE on my openSuSE 11.2 installation. I have never had any problems doing this in the past, but this time, when I rebooted, I noticed that after a few seconds of idle time, my CPU usage goes sky high. I ran top in a console and noticed the culprit was xorg. I am using an NVIDIA card on an AMD64 3200+ with 1 gig of memory. KDE version is currently 4.5.85. Like I said, I didn't have this problem until the last update. Any ideas as to what could be causing this?
I have recently upgraded from Kubuntu 9.10 to 10.04. I am having trouble with CPU usage by the X server. When running KDE programs (for example Kontact and Amarok) my computer is becoming very sluggish. This seems to be because X is using 90-100% of the CPU constantly when there are any changes to the application (any sort of redraw, even the mouse moving across the screen). I don't experience this with other applications (I am currently writing this from Chromium with no issues at all). Below is all the relevant information I can think to include.
INSTALL PROCESS: Fresh install from Ubuntu 10.04 LiveCD. My /home direrctory is on a separate partition so that has remained unchanged from 9.10. After installing Ubuntu, I installed the kubuntu-desktop package to install kdm and the KDE desktop.
The only real change from my last install is that /var is now also a separate partition (in a bid to save my MySQL data should I have to reinstall again) but it is on the same hard drive as my root partition and so shouldn't be causing any performance issues.
MY X SETUP: Three screen setup using onboard and PCI graphics cards. This worked perfectly before and, after reconfiguring X using nvidia-xconfig works almost as well again. Except that one of my monitors autodetects at 1280x1024 where before it was a higher (and widescreen) resolution.
The issue only presents itself when KDE application screens are visible and disappears when they are hidden (either minimised or closed) As far as Ican tell it could be an X, Qt or KDE issue. But I am not sure what versions of these will have been changed in the upgrade.
I don't understand why I have a 800MB usage at startup. I didn't add any programs to start at startup. I do though have 10 primary partitions that are auto mounted at startup but it shouldn't use up so much, I have 2GB total RAM.
At startup nautilus uses ~400MB of virtual memory in the "system monitor"
Also which is RAM and which is SWAP in resources in system monitor: Virtual Memory && Memory?
-i figure vm is ram but thats not possible sometimes...it exceeds 2GB
I play RuneScape most of the time i'm on the computer and i always keep an eye on the performance. Yesterday and the days before, everything was going good, cpu usage it was causing was between 30 and 60%, now it wont go below 70%. It's always 70-99% for some reason :s i have set settings for graphics to OpenGL (they were like that before too with no problem) and it started causing high cpu usage. If i set it to Safe Mode in graphics settings it gets 100%, I never tweaked the drivers, i have a NVIDIA GeForce 9600 GT, i have the opengl thingies installed along with my nvidia driver. Ubuntu 10.10 32bit.
I am calling clock_gettime() function to get time with nano seconds accuracy. My program works fine on Ubuntu but have high CPU usage problem on CentOS 5. it takes 40% CPU on Ubuntu and 90% CPU on CentOS. Kindly give me solution so that I can reduce high CPU usage problem on CentOS. You can build this code like: "gcc -lrt gettime.c -o gettime.e"