The aim of this script is, when the folder reaches 20M then attributes will be set to that particular folder so that no newfiles and folders cannot be created or copied to that samplefolder. whenever i copy a file morethan 20M to that folder its getting copied fully and then the attributes were applied. But i dont want this to happen, when the folder reaches its maximum current write operation to that folder should be stopped automatically with a error.
Is there any maximum limit to the heap memory allocation?My program is in PERL and i am using a solaris system. when i did "pmap pid" (pid = my process id) it is showing a number of heap memory allocated and all of them with GB sizes. This single process is eating up most of the physical memory.Is it normal and is there any way to get the heap memory size down
I don't know if this is related to a problem I have run into and posted elsewhere regarding 2 (or more) ethernet ports. But I encountered this while trying to solve the other problem (which I thought was Linux not setting correct routes for 2 interfaces). As suggested by someone when trying to solve the other problem, I switched to using the same IP address (it's a secondary address) on BOTH interfaces. So thus I have configured 172.30.0.13 on both eth0:1 and eth1:1. I am running the NSD program (an authoritative-only name server) listening on port 53 of 172.30.0.13. Some computers are getting the MAC address of eth0 for their ARP requests. Others are getting the MAC address of eth1 for their ARP requests. So this is determining which ethernet port their DNS queries will arrive on.
Those that send their DNS queries to the eth0 MAC address work fine. The NSD process gets the requests and answers them. The answers get back to where the query was sent from. HOWEVER ... those that send their DNS queries to the eth1 MAC address do not work. Using tcpdump, I see that the queries actually do arrive on the server. Using strace, I see that the NSD process never gets them. There are no iptables in effect.
Any idea why the kernel is deciding to not deliver the DNS query UDP datagram to the NSD process? It sure seems that the kernel just doesn't handle more than 1 ethernet interface (at least in the same subnet) correctly. IMHO, when an ARP request is received on 2 or more different interfaces, it should at least answer on both, each answer with the respective MAC address of that interface. It cannot know, and should not assume, any specific physical topology of the network beyond those interfaces.
When I try to delete a file, ( move to trash ) It says , The trash has reached It's maximum size! clean the trash manually. When I click on the trash icon on desktop it is empty. Where is the trash? Where can I delete these files ?
First off all, I'm booting from a large MEMDISK using PXE (900MB) . Due to our environment, I cannot decrease the size, nor move files to a nfs/iSCSI/... environment. Everything needs to be in that MEMDISK.
Now, when I try to run the OS, I get out of vmalloc space. How do I increase it to a number which allows such a large image to be mapped? I tried the parameter "vmalloc=1280M", but with that parameter, I don't get past the Booting the kernel screen.
Memory should not be an issue, since the machine(s) have at least 2GB RAM. (900MB MEMDISK + 256MB for other kernel stuff + 768MB for user stuff). The machine(s) have a Pentium 4 Extreme Edition processor, with hyperthreading and SSE2, but no EM64T.
How can I boot the system, and get past that message? Decreasing the MEMDISK size is not possible too. It is at the smallest we can get with our userland + kernel + modules.
I'm using fc14 and the SG driver to test some SCSI (SAS) targets. In doing so, I'm bumping up against what appears to be a 512KB maximum transfer size per command. Transfers up to 4MB sometimes work, but often they result in ENOMEM or EINVAL returned from the write() function in the SG driver. I could not find any good documentation on how the SCSI system in Linux works so I've been studying the source for drivers in drivers/scsi.
I see that there is a scsi_device struct that contains a request_queue struct that contains a queue_limits struct that contains an element called max_sectors. The SG driver seems to use this to limit the size of the reserve buffer it is willing to create. I see that there are several constants used to initialize max_sectors to 1024 which would result in the 512KB limit I see (with targets having 512 byte sectors). At this point I have several questions:
1) When the open() function for the sg driver gets called, who initializes the scsi_device struct with the default values?
2) Can I merely change the limits struct to arbitrary values after initialization and cause the SG ioctls to set the reserve buffer to allow greater values?......
Are there a way to set fixed maximum processing power to a process.For example a program called 'X' uses cpu in high speed. I need to set this 'X' process to to run with maximum 20% of cpu usage at any time.
I created a VM disk image with kvm-img, but I forget what was the max size of that disk image when I created it. Currently, its size is 6.2G, I want to install some large packages in that VM, so I want to make sure the disk image can expand to an adequate size.
I have set up squid server. My cache directory has been set up as per following statements.cache_dir ufs /Cache1/squid 10000 16 256cache_dir ufs /Cache2/squid 10000 16 256Now the problem is that size of /Cache1 and /Cache2 has reached to about 8GB and in near future it will reach the maximum limit of 10GB. I just want to know that whether I need to delete the contents of these directories or otherwise.
My requirement is nearly transferring 50GB files from remote server to local server via SCP command,while transferring the data its break-en because of some reasons in 35% itself. is it possible to resume the transfer process again from this 35% process.
This seems like a relatively simple question, but the answer seems to elude everyone: What is the MAXIMUM SIZE of a Linux loopback device (not counting any specific filesystem limitations)? Is it the maximum size of a linux block device?
I'm trying to build a multi-threaded server "C" program connecting to multiple clients. For various design reasons, I thought of having one port number per client. What is the maximum number of ports my process can handle
Make a new user Delete the "$HOME/Templates" folder Create a "$HOME/Templates" folder
Now, no new templates will be recognized in the right-click menu.
problems: Files created in the Templates folder are not found in the menu of <Right-Click> -> New Document.
In Nautilus, Go -> Templates sends me to /home/$USER/.
Restarting the computer does not solve this. I have done about an hour of searching on the internet and I have searched through the Nautilus source code.. I an not good enough with Linux to be able to understand the Nautilus source.
Firstly, I did perform a search on this problem in these forums, but didn't quite get what I was looking for. So I hope I don't yelled at for making a duplicate post. So I used rsync to backup my webroot to another nix machine. du -hs gave me 1.3 G on the source machine and 1.1 G on the backup machine. I tried to compare the individual files and noticed a trend. The files on backup machine were always smaller than the files on source machine. The source uses SATA drive, destination uses IDE. So this time I rsynced locally to another folder on the source machine. Same size anomaly. So i did a simple cp file ~/file and same size anomaly. So it's not a rsync issue.
I took a file and ran md5sum on both, the source file and destination file. To my surprise, even though the file size was different, they had the same md5sum. Now, let it be known that the source machine is a production server and the dir i rsynced was being used, serving pages to the web. I googled about this and came up with stuff like open descriptors and holes. I don't understand this stuff and was wondering if this was really the case. What are those if it is the case? And my backup copy is 100% identical right? There are thousands of files and I ran md5sum only on couple. Can I take comfort that when time comes, I can restore using my backup without any problems?
As I have a opt/sybase folder in Linxs I want to increase the size of this folder how can I increase it? For example the size we mention at the time of installation Sybase is 2gb now I want to increase by 4gb how can I increase it.
I installed Kubuntu 9.1 and then I downloaded Xampp of Linux to the desktop, however I have to extract it to /Opt/lampp. And that's the problem. I don't seem to have access or permissions to ?? I tried right-click | properties | permissions but the fields are greyed-out to make any changes.I am logged-in using the same user id/password I created when installing K9.1.