For Linux command sort, how do I force sort to load all input into memory and sort assuming I have enough memory? Or is it best to use a RAMDISK to store the input before feeding it to sort?
I'm running a java program and to debug the program instead of using a debugger I would like to look to the heap and to the bytes in the memory of the process. I was hoping that the pointers would give me a snapshot of a moment of the application. It exists a program that allow us to look to the details of a process and to the memory ram?
I'm still pretty new to the linux scene but kubuntu has been a great alternative to ubuntu (unity, ugh) until now.
It only started happening since i started messing around with play on linux but i dont think their related. I dont think its a driver issue either as ive switched between proprietary and default to no avail.
Basically after i log in, it takes over 3-6 minutes getting slower and slower before finally getting back to regular speed. So after a few occurences i decided that immediately after startup i went to system monitor and sure enough;
There is a process title "akonadi_contact" that is slowly growing in memory. There are also multiple occurences of it (usually between 6 and 9). they keep growing usually to the 500000k mark in memory usage before turning gray (swap storage??) and then eventualy the end.
What is "akonadi_contact" How do i stop this from happening. Is this a memory leak?
I have a simple C program that reads a 1 GB file from disk into memory.The first time I run this program, the process takes several seconds. But if I run this program again immediately after it finishes, the read happens almost instantly. What is going on here?My suspicion is that the operating system is realizing that the data I'm trying to read is still uncorrupted in memory (its been freed but not overwritten), so instead of reading it again, it just gives me the same block. I would really like to know what this process is called, or any Googleable keywords that would allow me to research it.
The reason I ask is because the size of my input data has increased. Now I need to read in a 6 GB file (my system has 12 GB of RAM), but I'm not observing the same behavior with the larger file. Each time I run the program the read takes an equally long time.
I am using ARM9(S3C2440) board in which linux kernel 2.6.30.4 is ported and i need to write some 8-bit data(say 0xAA) into the memory location 0x08000000 and i need to trace the data written into the location is correct or not how to do this in linux.
I have a mxn matric (which is my simplified way of saying it is RAM with bytes on it) Some of the locations on this metric is filled with some data and some places are empty. The mxn are very big numbers in size. I am trying to make a program so that if a system call wants to write some thing on empty locations on this mxn metric it should be able to do so without any problem. The thing which I want to understand or logic of a data structure is what data structure do you people feel should I be maintaining so that I can allocate the requested space immediately from the above mxn matric when some system call requests for some (k) number of locations from above metrics.
The logic initially I thought was to maintain a hashtable
1bytes requested----------> location 1,location 2,location 3.........location n 2bytes requested----------> location 1,location 2,location 3.........location n 3bytes requested----------> location 1,location 2,location 3.........location n
[code]....
but the problem with above logic is size of the pointers where I will be writing this problem is unsigned 64 byte.So to know location of one free byte if I am maintaining one pointer of type u64 this is not a feasible solution.
I have very little linux experience. And need some help with a bash script. I need to a script I can set cron to run to sort files out of a holding folder into final folders. It doesn't necessarily have to be bash, but I think it would be sufficient for this. File names are formatted as such when created: Dest-Date-Time-CID-Destination# I want the files to be moved from a all in one holding folder to a folder structure like this.
So the script will need to make directories based on information in the file name which is delimited by single dashes. Then move files from the holding folder to the newly created "sorted" folders.
A friend of mine asked me to try and recover some photos she deleted by accident on her SD memory card. Normally, I would flip over to my Windows box and do this, but I'm wondering if there is a Linux equivalent to a Windows data recovery program. I'm running Fedora 10.
I stupidly unplugget my USB-cable, which was connected to my Nokia music phone, just as if I were in Windows. What do I do? I've lost my music on the phone, or, it seems it may be there (the correct mass of data), but my phone now tells me there is no music... Can I recover this? And - what is the correct way to unplug a USB unit in Ubuntu? To make it work, and find the phone/drive - I just typed "sudo lsusb" in the terminal, and it found and opened the memory automatically... How should you unmount the USB, and maybe even how do I get my data back?
I am using malloc and frees a lot in my program. It shows its allocated but when i remove it doesnt show as the memory is removed(I am using the top command to view VIRT memory usage). If this continously grows what would happen to my program (Will it go out of memory?)
I have a computer with 16GB of ram. At the moment, top shows all the RAM is taken, (NOT by cache), but the RAM used by the various processes is very far from 16GB.I have seen this problem several times, but I don't understand what is happening.My only remedy so far has been to reboot the machine.
I have just started to learn Java programming and need to be able to use a compatible notepad equivalent in Ubuntu i will need to save the data as is,in Ubuntu and be able to transfer the data from my Linux machine to the windows desktop regularly via memory stick i absolutely love Linux and don't want to have to go back to windows to do this.
This is my first post in these forums. I'm still quite new to Linux (using Mint 9) so please bear with my not-very-articulate question(s)When I boot up and open up a tty terminal I get a message saying "Memory corruption detected in low memory." I've done an extensive google search about the issue and it seems not uncommon. I ran a memtest with no errors returned, so I'm sure that there's nothing really wrong with the memory; apparently it's a bug in the kernel that's causing this.
I found from command 'top' that 8GB memory are used. However, using command 'ps' with some options to grep the running processes and then summing up the memory used by the running processes are less than 2 GB. Where has the used memory gone ?
Last week my server crashed. I'm trying to diagnose the cause.
This is the relevant error message in /var/log/messages:
Code:
I'm assuming that I can conclude, then, that apache/httpd was the cause of the memory leak?
Next, I've been tracking my memory usage. Using top, this is an average memory load level for my server:
Code:
I'd like to confirm if my understanding of this data is correct, because Plesk indicates that my memory usage is only 50% or less. (Though I have read a number of reports indicating that Plesk's measurements are frequently wrong.)
Top says: Of the 2,073,156K total memory, 1,982,572K (95.63%) is being used, 90,584K (4.37%) is free. Of that sum, 421,948K (20.35%) are being used as buffers. Additionally, of the 2,096,472K of Swap, 60K is used, and 887,700K (42.34%) is cached.
My questions: Is my memory actually being 95% used? Or is the buffered quantity (20.35%) not a use of physical/virtual memory? (i.e. is it disk usage?) Does the amount of cached Swap influence the percentage of physical/virtual memory being used?
In other words, who is correct? Plesk says I'm using 40-50% of my memory, whereas top says 85-95%.
The current Octave 3.4.0 on Fedora 15 is giving me an "out of memory" message when I plot certain data. See example below. It depends on the range, and the particular function. Anyone else get this behaviour?
A few months ago I have setup a server with three hard disks. The partition mapping the disks as follows:
Disk /dev/sda: 1000.2 GB, 1000204886016 bytes 255 heads, 63 sectors/track, 121601 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Disk identifier: 0x7ca36fee
[code]....
Now I have the following problem the LVM file system don't mount properly.If I open the mount point I see only a few files of the LVM disk. If I want to unmount the disk I get the following error:
umount /data/ umount: /data/: not mounted
If I want to mount the volume I get the following error:
mount -a mount: /dev/mapper/gegevens-Data already mounted or /data busy
What options should I use when I'm using the sort command to sort the top 5 CPU processes (ps -eo user,pid,ppid,%cpu,%mem,fname | sort ??? | head -5) showing max to min usage?
We switched from unix to linux and we have an old report that extracted data from a database, output to an ascii file and then sorted the results in the file based on different arguments. The report now blows up when it runs,and I can only guess it is because the options for sort on linux differ slightly from unix.For example, here is one of the commands issued from within the report app that ran on the old unix box:
I will eventually rewrite the report to store the data in a local table, but I can simply adjust the options to suit the requirments of linux. Basically, I need to know if this can be a quick fix for the short term.
I am looking to buy some memory for my netbook. Currently I have 1 GB of DDR3 memory. However, the specification says that 2 GB of memory is the max. However, when I do the following it says that 4GB is the max:
I am looking for free database that has low memory usage and innodb and memory like engins that has C API and support trigger and client/server support for using in embedded linux systems.
I am new to C and linux. My code below does arbitary writes but I cant figure out where or how it does it.
I am calling the insertNode() function with seq = 'MISSISSPPI$' and alphabets = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ$'
Code:
Weird behaviour I should mention is that when I check for NULL pointer in node->child[index], the unassigned values are not null anymore, they point to arbitary memory.
we found that if we use 'top' to show the memory usage of a server (SuSe Linux 10), we can get virtual memory usage as well as 'Resident memory' usage. For virtual mem or a particular process, it is around 1.1GB, which is large but for resident memory, it only consumes 300MB. Are there anyone who knows what the differences are? I would also like to know whether the difference (1.1GB - 300MB) = 800MB are actually available for use by other applications in the system.
I am monitoring physical memory in a server I administer, and my hardware provider told me they had increased physical memory size to 4Gb... However, using several tools (free -m; top; dmesg | grep Memory; grep MemTotal /proc/meminfo I discovered that I actually have 3Gb, not 4... But, my doubt comes from the fact that dmesg | grem Memory tells me I have 3103396k/4194304k available The first number is effectively 3Gb, but the second one, is 4! so, why I am looking at this two different numbers?