General :: Memory Leak Whenever Try To Connect The Server Using The ()?
Jun 3, 2010I am facing problem of Memory leak whenever i try to connect the Server using the ().getting rid of this problem.
View 1 RepliesI am facing problem of Memory leak whenever i try to connect the Server using the ().getting rid of this problem.
View 1 RepliesI am facing problem of Memory leak whenever i try to connect the Server using the ().
View 2 Replies View RelatedI'm afraid I have a huge issue with my newest Fedora 14 server. I recently migrated to Fedora 14 from Centos 5, which was very stable, but had ancient packages and libraries and my users were revolting...The machine is a HP ProLiant 360 G7, with 12G RAM and 6 SAS drives in RAID 5.After I migrated to Fedora 14, I noticed that for some reason, during the course of about 24 hours, all usable RAM "disappears" and applications are forced down to swap space. Needless to say I didn't have this issue on CentOS.
The server does heavy IO as per it's function (it's a heavily loaded file processing server and user simulation computing station among other things, which causes lots of random IO), so I thought it may be the cache, but then I realized it cannot be - because obviously Linux will use onyl "unused" RAM for caching and frees it up as soon as an app need it. Then, I thought to check the "slabtop" to see what's going on in Kernel memory. Unfortunately I don't have the screenshot from the time just before the latest crash, but there's a certain value displayed by slabtop, which slowly, byte-from-byte creeps over all available RAM, eventually forcing applications down to the swap. This is malloc-64, and as you can see from the bellow copy-paste, it's building up again even now...
Code:
Active / Total Objects (% used) : 9118075 / 9153600 (99.6%)
Active / Total Slabs (% used) : 152157 / 152157 (100.0%)
[code]...
My server [URL] goes down in some cases.
On general time load average is 0,15-0,2 Only http(with php) and mysql working on it.
Here are some screens when the server goes down: [URL]
Here are the same bug with mine described: [URL]
Is there any link where i can get information about below?
Dirty memory
RSS
PSS
One more?
if a set of process are getting executed in a use case say 50 times. How do one know the memory leak for a particular process?
I have had a fresh install of Ubuntu 9.10 and installed some software after that.Since third some, some process is eating half of my memory.I have checked processes running in system manager but everything is normal.Maximum is consumed by compiz which is about 26 mb, seems very normal.I did restarted my computer several times, and in the start for 5 mins, its fine after that again my cpu fans runs at very fast speed and my one cpu is used up 95 % (I have dual core).Please help me out, this invisible thing is driving me crazy.I am attaching my htop screen shot (sorted by cpu %), now the cpu is not used by completely but fan is still struggling hard and fast.
View 9 Replies View RelatedI'm currently developping a program in C++, using Qt, for an embedded board (SBC9261).It works well but crashes after some time, due to a system memory overload (my program uses more and more memory until 100%, when it crashes).I've been able to figure out the source of the memory leak :The function f is called by my program every second. f instanciates a new object (a QImage from the Qt lib), does a bunch of processing on it, and returns it to the calling function :
Code:
QImage *MyClass::f(QString filename) {// Open image
QImage *image = new QImage(filename);
[code]....
root@XXXXX:~# uname -a Linux myserver 2.6.18-194.11.4.el5 #1 SMP Tue Sep 21 05:40:24 EDT 2010 x86_64 x86_64 x86_64 GNU/Linux I'm having serious memory leak issues on a server running CentOS. Running 'top' I can't find any proccess with an unusual memory usage. Is there any other way to check what is using this memory? Right now it shows that I have 4.8GB of RAM used, but top shows only few proccesses, one with 4% and lots of 0.0%.
View 9 Replies View RelatedI need to write a small program that eats away at availabe memory. I need to creaty a memory leak to test how other programs cope. I need to run this program on linux and see if the available memory is decreasing.
So I have done:
Code:
int main()
{
int *buffer;
while(1)
[Code]...
I have 4 gigs of ram and xorg goobles up 70+% in a matter of an hour or so. I'm running a java app with lots of 2D graphics, firefox, pidgin, and one other java app. I recently upgraded to KDE 4.2 and got dual ATI firepro 3700 cards running 3 22" widescreens. The memory gets eaten until my java app crashes.
I'm using the fglrx driver as it seems the radeonhd one doesn't recognize the firepro 3700. Can one only have a single KDE4 version on the machine. I would like to be able to swtch back and forth between 4.1 and 4.2 in order to troubleshoot without all the downloading and installing in yast.
I have encountered this problem of memory usage is increasing as the during the my program is being run until 1Mb is left then it stays at that.A part of the program is this:
Code:
#define WRITE_BUF_SIZE (1024*1024)
void post_data(const void *data, unsigned long size)
[code]....
I can confirm my machine's GLX version is 1.4, and I definitively get the slow down problem, however, after running in a terminal:
Code:
grep "object bytes" /sys/kernel/debug/dri/0/gem_objects
grep: /sys/kernel/debug/dri/0/gem_objects: No such file or directory
What am I doing wrong?
I dont know if this issue has already been solved, or if there are other threads dealing with it, but Im quite desperate with this bug that has been annoying me since I upgraded from 9.10 to 10.04.
Firs I thought it would be GEM related, but Im using a proprietary graphics card, so thats not the way to solve it. I experience the problems a memory leak provokes, mainly that, as time goes by with the laptop turned on, the system needs more resources to execute simple actions, and it freezes too many times. Plus IBUS will, suddenly, stop working (I still don't know if it is related or not), and, most of the times I log in, Notification Area 2.30.0 will load with errors.
I'm having some strange issues with apache. Time by time it segfaults, eats all available memory (including swap) and makes server non responding.Ubuntu Server 10.04.2 LTS
Some strange logs:
Jun 12 12:00:18 *** kernel: [40767.969443] apache2[7635]: segfault at 726f7272 ip 00007f13a31f3f16 sp 00007fff6f740ea0 error 4 in libapr-1.so.0.3.8[7f13a31d7000+35000]
[code]....
I installed Ubuntu 10.4 and everything works great "out of the box"! I didn't install any drivers.blem, a major one...When I type "top" right after a reboot, around 1GB of RAM is in use. After a few min it grows to 2GB, even 2.5GB for no reason. Luckily, my machine has 6GB of RAM but it's still a major issue for me.I read on another forum another user with the same problem on the same Lenovo machine and he solved it by installing the newest driver for his ATI device. I don't have ATI, but I have the impression that I can solve my problem the same way. Do you have any idea what driver my cause that? I checked the Intel site for a new Intel HD graphics driver but they don't have ones for Linux.
View 2 Replies View RelatedI updated several packages on one of my servers on Dec 21st and have been seeing excessive swapfile usage since then. The problem process seems to be httpd which in our environment runs a subversion server as well as serving a number of php pages over https. At present I am having to bounce apache approximately every 5 days as it has used all 8GB swap in that time.
[Code]..
Of the updates listed as installed, the only one that looks likely to affect apache is glibc. Looking at the stats from sar -r I can see swap usage increasing by approx 3% (of 8GB) every hour.
I'm using Fedora 15 with Gnome 3 on a 32 bit laptop. I noticed that there seems to be a huge memory leak issue with Gnome shell. When I restart it, it is around 20 MB. But it keeps rising, and after around eight hours, I noticed it was around 250 MB! I found a solution online that said to simply restart the shell if the gnome-shell memory consumption becomes too large. While this is fine as a temporary solution, I am looking for a permanent one. Is there a way to minimize/prevent the memory leak other than waiting for the next version of Gnome 3?
View 3 Replies View Relatedlet me know how to clear cache memory ( RHEL 5.1 ) as it consumes almost 100% physical memory.
View 3 Replies View Relatedwe found that if we use 'top' to show the memory usage of a server (SuSe Linux 10), we can get virtual memory usage as well as 'Resident memory' usage. For virtual mem or a particular process, it is around 1.1GB, which is large but for resident memory, it only consumes 300MB. Are there anyone who knows what the differences are? I would also like to know whether the difference (1.1GB - 300MB) = 800MB are actually available for use by other applications in the system.
View 1 Replies View RelatedI have been setting up a vps I got out with bhost.net, with CentOS installed. I've been learning and have set up everying I need with the exception of ftp/sftp.
Using yum I installed vsftpd and ran into problems, thinking it was something I might of done I did a fresh install of CentOS and I still recieve the same problem on a fresh install so it is nothing I have done to the server.
The problem is when connecting via a sftp client I get an out of memory error. This error is listed in the putty faq ( url ) under A.7.5, there is a brief explaintion of the cure under A.7.6.
there is mention of a login script but I don't know where this is located. I'm a novice at Linux but by no means incompotent when it comes to computing.
how can i find out how much memory my server has?
i am running centOS 5
I need to check if my DIMMs are working fine on my server.
View 4 Replies View RelatedI am using malloc and frees a lot in my program. It shows its allocated but when i remove it doesnt show as the memory is removed(I am using the top command to view VIRT memory usage). If this continously grows what would happen to my program (Will it go out of memory?)
View 4 Replies View RelatedI have a computer with 16GB of ram. At the moment, top shows all the RAM is taken, (NOT by cache), but the RAM used by the various processes is very far from 16GB.I have seen this problem several times, but I don't understand what is happening.My only remedy so far has been to reboot the machine.
View 1 Replies View RelatedI have a query regarding top & virtual memory. When we run top it show VIRT (Virtual Mem), RES (Resident Mem) & SHR (Shared Memory). The total virtual memory of my machine is 4 GBs (2 GB RAM + 2 GB Swap), but still I am able to see a process showing 4000m virtual memory column. what it means, as its show VIRT Mem more than actual available VIRT memory
View 1 Replies View RelatedThis is my first post in these forums. I'm still quite new to Linux (using Mint 9) so please bear with my not-very-articulate question(s)When I boot up and open up a tty terminal I get a message saying "Memory corruption detected in low memory." I've done an extensive google search about the issue and it seems not uncommon. I ran a memtest with no errors returned, so I'm sure that there's nothing really wrong with the memory; apparently it's a bug in the kernel that's causing this.
View 2 Replies View RelatedI found from command 'top' that 8GB memory are used. However, using command 'ps' with some options to grep the running processes and then summing up the memory used by the running processes are less than 2 GB. Where has the used memory gone ?
View 7 Replies View RelatedI am implementing an automated backup scheme so I created a shell script which first creates SQL dumps for all MySQL databases, then retrieves all websites from the /var/www directory of a remote server. The latter is working as I am using rsync to get the remote files.However, the MySQL dumps being retrieved are the ones from the local server which is not what I want. I want to get the SQL dumps from the remote server as well.I have a tunnel between the local and remote server which I can connect to without using any password (I added the public key to the authorized_hosts), so I tried to add the following code to the script:
ssh user@192.168.x.xxx
I then attempted to retrieve the SQL dumps and exited from the remote server. However this does not work as I still have to enter exit manually in the terminal for the SQL dumps to be retrieved from the remote host. I don't know why this is happening. This is what the script is trying to do:
//connect to remote server
ssh user@192.168.x.xxx
//retrieve SQL dumps
[code]...
Is there a way to connect to the remote host AND run the script's code on THAT remote host?
When using TOR it is recommended that you disable javascript and all other scripts because they can be used to detect your IP, defeating the purpose of TOR.
But there are some sites that simply do not work without javascript. Is it possible to keep all javascript functionality but disable just the functionality that is used to leak your IP?
I created an account called adminftp on a local Linux server Centos.Added this user to FTP group.When I use Filzilla from another computer in the same network, it cannot connect to the server but I can ping to the server.
View 6 Replies View Related