While Apache is off, then everything is OK, but if I start apache server after some time I get messages "Unable to fork: Cannot allocate memory" when trying to do some commands with ssh(dir,rm,top and so on).
I recently had to move to a new machine, everything went well except for one thing. I did fresh installation of LAMP server all with default configs. Every time I'm using PHP script to that invokes include, require or require_once I get the following error:
Code: Fatal error: Allowed memory size of 20971520 bytes exhausted (tried to allocate 7680 bytes) in /var/www/index.php on line 2 index.php file: PHP Code: <?php include "index.php";?> icukapi.php file: PHP Code: <?php echo "test";?>
My memory_limit in php.ini is set to 20M. I tried to increase that however it didn't quite work. PHP seems to allocate all possible space and return that message every time i try. If somebody has an idea of how to fix it I would be more than grateful. I spend quite a long time searching for an answer however the all things i found suggested increating memory_limit which in this case doesn't work.
i just recently installed ubuntu 10.04 onto my laptop. its running as a dual boot. i only allocated about 10gb of memory and have already run out. is it possible to allocate more memory afterwards
and another bug that i am having is my shift key is not functioning properly. if anyone happens to know this issue too. so i can't do exclamations of periods, capitalizations, etc... how would you remap your keyboard
Do notice the line in the above quote ""sh": java.io.IOException: error=12, Cannot allocate memory."
The server will generate a world, but you can not connect to it using the MC client. Does anyone know what the problem is? I have googled for this problem for literally the past 4 hours and I am getting no where.
I am running BWA (burrow-wheeler alignment) and, keep getting an error that there isn't enough memory allocated. How do I change the memory allocation? I'm working on a good quality desktop only a little over a year old, so having the memory shouldn't be the problem.
I'm running a quite complex model, and it's taking me around a week to complete the run, and one of my colleagues says that he can run it in one-two days. After looking to several forums to know how does it work, I've seen how the memory is used (using the command ps -A --sort -rss -o comm,pme), and the model run just takes 0,3% of the memory. I bought a 4 cores and 8gb of RAM laptop to be able to run the model fast, but I don't know how to allocate the use of the memory to concentrate it in developing that task. Is that possible?
Can anyone tell me is there a way to add more than 2GB memory application in a 32 bit server? I found on many threads that it is not possible to set memory limit of a particular application to more than 2GB. I am using Fedora release 8 and trying to add more than 2GB memory to tomcat. If I modify the variable CATALINA_OPTS in /opt/bobcat/bin/catalina.sh (Startup script of tomcat) and add a value greater than 2GB for memory, tomcat wont start.Is there any work around / Kernel Patches available? Suggestions??
I went to an interview last week and there was this guy who asked a simple question that i have been trying to solve for a couple of days. I tried google but i just cant get the search keywords right. The result were just useless. Well, the question is : "How can we allocate a limited memory to a process before we start its execution" well, the question is related to an X11 system so may be some flags must be set to limit its memory.
I am developing a code where I need to store the planes of an object in a tree & also i need to store it in a list for further processing ... when I try to allocate using malloc the memory allocation when i checked this in internet , I came to know that it may because the memory that I am trying to allocate may be more than the size_t variable.
I just switched over to openSUSE about a week ago from Ubuntu and I just purchased an Asus N13 PCI-E wireless network adapter which uses an RALink 2860 chipset. I plugged it in and openSUSE was able to detect it and everything but when I went to go configure it using YaST, it wouldn't be able to scan for networks.
So, I reinstalled my old adapter, downloaded the latest source from RALink and compiled the drivers. When I tried to use insmod on the new .ko that was compiled, I got the "Could not allocate memory" error.
If you need dmsg logs or anything else, I can reinstall the card and get those.
I have been setting up a vps I got out with bhost.net, with CentOS installed. I've been learning and have set up everying I need with the exception of ftp/sftp.
Using yum I installed vsftpd and ran into problems, thinking it was something I might of done I did a fresh install of CentOS and I still recieve the same problem on a fresh install so it is nothing I have done to the server.
The problem is when connecting via a sftp client I get an out of memory error. This error is listed in the putty faq ( url ) under A.7.5, there is a brief explaintion of the cure under A.7.6.
there is mention of a login script but I don't know where this is located. I'm a novice at Linux but by no means incompotent when it comes to computing.
On my windows console, i have installed sun virtualbox. I only have 768MB ram and so i downloaded a trial version of eboostr. I thought it would work, but when i tried an install of 11.2, also, i tried a debian install, it did the same thing, it froze up in the middle of the install with the error Code: unable to allocate and lock memory. I guess that eboostr doesn't work, but is there a way to use a flash drive as ram in virtual box?
After updating to openSUSE 11.3, and restarting my computer, I placed my computer in sleep(suspend to RAM) mode, and I lost my internet connectivity. I noticed that my computer was only connected through ethernet. After entering "dhclient eth0" in a terminal, I received the following output.
RTNETLINK answers: Cannot allocate memory
This is the first time that I have had issues with connecting to the Internet using Ethernet.
My mail server stopped working this weekend, when I got in today I restarted and all the emails came through. The panic_log shows: daemon: accept process fork failed: Cannot allocate memory several times over the course of a few minutes. Can anyone give me advice on troubleshooting this one?
I am writing a function(in C language in Linux) to initialize an array of a structure from a mysql table by calling a function by reference but I have a problem with call this function by reference . this is my program:PHP Code:
I have a computer with 16GB of ram. At the moment, top shows all the RAM is taken, (NOT by cache), but the RAM used by the various processes is very far from 16GB.I have seen this problem several times, but I don't understand what is happening.My only remedy so far has been to reboot the machine.
I am using malloc and frees a lot in my program. It shows its allocated but when i remove it doesnt show as the memory is removed(I am using the top command to view VIRT memory usage). If this continously grows what would happen to my program (Will it go out of memory?)
I am looking to buy some memory for my netbook. Currently I have 1 GB of DDR3 memory. However, the specification says that 2 GB of memory is the max. However, when I do the following it says that 4GB is the max:
I have had a fresh install of Ubuntu 9.10 and installed some software after that.Since third some, some process is eating half of my memory.I have checked processes running in system manager but everything is normal.Maximum is consumed by compiz which is about 26 mb, seems very normal.I did restarted my computer several times, and in the start for 5 mins, its fine after that again my cpu fans runs at very fast speed and my one cpu is used up 95 % (I have dual core).Please help me out, this invisible thing is driving me crazy.I am attaching my htop screen shot (sorted by cpu %), now the cpu is not used by completely but fan is still struggling hard and fast.
I am monitoring physical memory in a server I administer, and my hardware provider told me they had increased physical memory size to 4Gb... However, using several tools (free -m; top; dmesg | grep Memory; grep MemTotal /proc/meminfo I discovered that I actually have 3Gb, not 4... But, my doubt comes from the fact that dmesg | grem Memory tells me I have 3103396k/4194304k available The first number is effectively 3Gb, but the second one, is 4! so, why I am looking at this two different numbers?
we found that if we use 'top' to show the memory usage of a server (SuSe Linux 10), we can get virtual memory usage as well as 'Resident memory' usage. For virtual mem or a particular process, it is around 1.1GB, which is large but for resident memory, it only consumes 300MB. Are there anyone who knows what the differences are? I would also like to know whether the difference (1.1GB - 300MB) = 800MB are actually available for use by other applications in the system.
I have to upgrade memory in a couple of remote (other part of the city) machines and I would like to be able to query all necessary info in a comfortable ssh session without having to drive there and unsrew each different machine (many different types). I googled a lot but I can only find Windows tools, nothing for Linux.
- total number of memory slots on the mainboard
- for each used slot: current module type, memory type, size and speed
so I can make a shopping list for all the different needed memory modules.
When using the Universal USB Installer and following the instructions according to the Ubuntu web site, I get the following error: an error () "occurred while executing syslinux. Your USB drive won't be bootable" There is no number between the brackets and I has no problem with a previous version of Ubuntu netbook remix.
Is that possible that SHM shared memory is counted as cache memory on Linux with kernel 2.6.18?If find it really odd since this memory is not file backed, but I have a piece of code that loads data using shm_open+mmap, and it generates an amount of cache memory in /proc/meminfo that corresponds exactly to the amount of shared memory (I load that data from a file but I am using posix_fadvise(fd,0,0,POSIX_FADV_DONTNEED) to ensure this file is not cached and I made sure that it is working as expected). As far as I know SHM memory was not tagged as cache memory with kernel 2.6.9.If it is the case it is really unfortunate since normally cache memory can be considered to be part of the "available" memory since it can be flushed promptly but this is clearly not the case with SHM memory... Is there an easy way to get the total amount of used SHM memory on a system?
I found from command 'top' that 8GB memory are used. However, using command 'ps' with some options to grep the running processes and then summing up the memory used by the running processes are less than 2 GB. Where has the used memory gone ?
This is my first post in these forums. I'm still quite new to Linux (using Mint 9) so please bear with my not-very-articulate question(s)When I boot up and open up a tty terminal I get a message saying "Memory corruption detected in low memory." I've done an extensive google search about the issue and it seems not uncommon. I ran a memtest with no errors returned, so I'm sure that there's nothing really wrong with the memory; apparently it's a bug in the kernel that's causing this.