General :: Break Copy Process When Folder Size Reaches Its Maximum?
Mar 19, 2011
I have written a script.
#/bin/bash
a=`du -sh /root/samplefolder | cut -f1`
echo $a > testfile
[code]....
The aim of this script is, when the folder reaches 20M then attributes will be set to that particular folder so that no newfiles and folders cannot be created or copied to that samplefolder. whenever i copy a file morethan 20M to that folder its getting copied fully and then the attributes were applied. But i dont want this to happen, when the folder reaches its maximum current write operation to that folder should be stopped automatically with a error.
View 5 Replies
ADVERTISEMENT
Aug 4, 2011
how to create a script when my mailqueue (mqueue) folder size reaches 1GB and sent email alert to my email id.
Once mqueue folder size reaches the sent alert by mail. Because everytime my mqueue folder size increasing.
View 3 Replies
View Related
Mar 15, 2011
Is there any maximum limit to the heap memory allocation?My program is in PERL and i am using a solaris system. when i did "pmap pid" (pid = my process id) it is showing a number of heap memory allocated and all of them with GB sizes. This single process is eating up most of the physical memory.Is it normal and is there any way to get the heap memory size down
View 3 Replies
View Related
Mar 30, 2010
I'm wondering what's the maximum heap space a process can use (not necessarily by a single malloc()) in Ubuntu x86_64. Which parameter determines the size
View 7 Replies
View Related
Apr 14, 2010
I don't know if this is related to a problem I have run into and posted elsewhere regarding 2 (or more) ethernet ports. But I encountered this while trying to solve the other problem (which I thought was Linux not setting correct routes for 2 interfaces). As suggested by someone when trying to solve the other problem, I switched to using the same IP address (it's a secondary address) on BOTH interfaces. So thus I have configured 172.30.0.13 on both eth0:1 and eth1:1. I am running the NSD program (an authoritative-only name server) listening on port 53 of 172.30.0.13. Some computers are getting the MAC address of eth0 for their ARP requests. Others are getting the MAC address of eth1 for their ARP requests. So this is determining which ethernet port their DNS queries will arrive on.
Those that send their DNS queries to the eth0 MAC address work fine. The NSD process gets the requests and answers them. The answers get back to where the query was sent from. HOWEVER ... those that send their DNS queries to the eth1 MAC address do not work. Using tcpdump, I see that the queries actually do arrive on the server. Using strace, I see that the NSD process never gets them. There are no iptables in effect.
Any idea why the kernel is deciding to not deliver the DNS query UDP datagram to the NSD process? It sure seems that the kernel just doesn't handle more than 1 ethernet interface (at least in the same subnet) correctly. IMHO, when an ARP request is received on 2 or more different interfaces, it should at least answer on both, each answer with the respective MAC address of that interface. It cannot know, and should not assume, any specific physical topology of the network beyond those interfaces.
View 1 Replies
View Related
Mar 8, 2011
When I try to delete a file, ( move to trash ) It says , The trash has reached It's maximum size! clean the trash manually. When I click on the trash icon on desktop it is empty. Where is the trash? Where can I delete these files ?
View 2 Replies
View Related
May 31, 2011
First off all, I'm booting from a large MEMDISK using PXE (900MB) . Due to our environment, I cannot decrease the size, nor move files to a nfs/iSCSI/... environment. Everything needs to be in that MEMDISK.
Now, when I try to run the OS, I get out of vmalloc space. How do I increase it to a number which allows such a large image to be mapped? I tried the parameter "vmalloc=1280M", but with that parameter, I don't get past the Booting the kernel screen.
Memory should not be an issue, since the machine(s) have at least 2GB RAM. (900MB MEMDISK + 256MB for other kernel stuff + 768MB for user stuff). The machine(s) have a Pentium 4 Extreme Edition processor, with hyperthreading and SSE2, but no EM64T.
How can I boot the system, and get past that message? Decreasing the MEMDISK size is not possible too. It is at the smallest we can get with our userland + kernel + modules.
View 1 Replies
View Related
Jul 17, 2011
How do I change the size of the available shared memory on Linux?evidently 4GB is not enough for what I am doing (I need to load a lot of data into shared memory - my machine got 8GB of RAM).
View 2 Replies
View Related
Apr 7, 2011
I'm using fc14 and the SG driver to test some SCSI (SAS) targets. In doing so, I'm bumping up against what appears to be a 512KB maximum transfer size per command. Transfers up to 4MB sometimes work, but often they result in ENOMEM or EINVAL returned from the write() function in the SG driver. I could not find any good documentation on how the SCSI system in Linux works so I've been studying the source for drivers in drivers/scsi.
I see that there is a scsi_device struct that contains a request_queue struct that contains a queue_limits struct that contains an element called max_sectors. The SG driver seems to use this to limit the size of the reserve buffer it is willing to create. I see that there are several constants used to initialize max_sectors to 1024 which would result in the 512KB limit I see (with targets having 512 byte sectors). At this point I have several questions:
1) When the open() function for the sg driver gets called, who initializes the scsi_device struct with the default values?
2) Can I merely change the limits struct to arbitrary values after initialization and cause the SG ioctls to set the reserve buffer to allow greater values?......
View 2 Replies
View Related
Jan 31, 2010
Are there a way to set fixed maximum processing power to a process.For example a program called 'X' uses cpu in high speed. I need to set this 'X' process to to run with maximum 20% of cpu usage at any time.
View 4 Replies
View Related
May 22, 2011
I created a VM disk image with kvm-img, but I forget what was the max size of that disk image when I created it. Currently, its size is 6.2G, I want to install some large packages in that VM, so I want to make sure the disk image can expand to an adequate size.
View 1 Replies
View Related
Jan 20, 2010
I have set up squid server. My cache directory has been set up as per following statements.cache_dir ufs /Cache1/squid 10000 16 256cache_dir ufs /Cache2/squid 10000 16 256Now the problem is that size of /Cache1 and /Cache2 has reached to about 8GB and in near future it will reach the maximum limit of 10GB. I just want to know that whether I need to delete the contents of these directories or otherwise.
View 1 Replies
View Related
Apr 28, 2011
I would like to ask you if there is any maximum allowed number of files per folder in linux (without risking it to lose everything). I am using openuse 11.4 with latest kde (4.6?).
I am trying something fast and dirty and it might be that one folder will contain like 10^6 files.
Is there is anything I should be warned about that?
View 10 Replies
View Related
Nov 23, 2010
In Linux bash shell, for a given directory, how can I list:The create date for that directory The number of files in that directory The number of subdirectories in that directory.
View 3 Replies
View Related
Jul 2, 2010
Is there a way to recreate all the folders from one directory to another without copying over the contents of the folder? I've been trying to do something like this,
Code:for i in `ls $X`; do mkdir $PATH/$i; doneUnfortunately $i is deliminated by whitespaces in the filenames and not the actual folders.
$X contains only other folders so I dont have to worry about regular files but any kind of more "advanced" solution would work.
View 1 Replies
View Related
Jul 23, 2011
My requirement is nearly transferring 50GB files from remote server to local server via SCP command,while transferring the data its break-en because of some reasons in 35% itself. is it possible to resume the transfer process again from this 35% process.
View 3 Replies
View Related
Mar 4, 2010
i need to copy files from a folder called output files to another folder called running but im not able to do it i always get a stat error
View 14 Replies
View Related
Feb 2, 2010
how to mkpart with maximum size available?
mkpart primary 0 3001G
I have 3TB raid disk. I cannot install centos using GPT partition. I need to use parted and mkpart to format it.
View 1 Replies
View Related
Jun 4, 2011
I use U-server. I set up an mail server as described in Ubuntu Documentation > Ubuntu 11.04 > Ubuntu Server Guide > Email Services.
How can I turn off maximum attachment size. On my server I don't want to limit the size of e mails that arrives.
I manage my server with webmin
View 2 Replies
View Related
Sep 24, 2010
This seems like a relatively simple question, but the answer seems to elude everyone: What is the MAXIMUM SIZE of a Linux loopback device (not counting any specific filesystem limitations)? Is it the maximum size of a linux block device?
View 4 Replies
View Related
Nov 28, 2009
How do I find out maximum I/O request size?
I'm using CentOS release 5.2
I know that the default size is 128 K, how do I find out/confirm this?
View 1 Replies
View Related
Apr 2, 2010
How can I maximum the CPU usage for a process? And also use both cores if possible.
View 5 Replies
View Related
Dec 8, 2010
I'm trying to build a multi-threaded server "C" program connecting to multiple clients. For various design reasons, I thought of having one port number per client. What is the maximum number of ports my process can handle
View 5 Replies
View Related
Mar 10, 2011
I am using Nautilus To re-create the problem:
Make a new user
Delete the "$HOME/Templates" folder
Create a "$HOME/Templates" folder
Now, no new templates will be recognized in the right-click menu.
problems: Files created in the Templates folder are not found in the menu of <Right-Click> -> New Document.
In Nautilus, Go -> Templates sends me to /home/$USER/.
Non-solutions:
Restarting the computer does not solve this. I have done about an hour of searching on the internet and I have searched through the Nautilus source code.. I an not good enough with Linux to be able to understand the Nautilus source.
View 3 Replies
View Related
Feb 9, 2010
How to restrict the maximum size of an array variable in perl such that Ishould not be able to have array elements beyond the maximum size?
View 7 Replies
View Related
Feb 27, 2010
Firstly, I did perform a search on this problem in these forums, but didn't quite get what I was looking for. So I hope I don't yelled at for making a duplicate post. So I used rsync to backup my webroot to another nix machine. du -hs gave me 1.3 G on the source machine and 1.1 G on the backup machine. I tried to compare the individual files and noticed a trend. The files on backup machine were always smaller than the files on source machine. The source uses SATA drive, destination uses IDE. So this time I rsynced locally to another folder on the source machine. Same size anomaly. So i did a simple cp file ~/file and same size anomaly. So it's not a rsync issue.
I took a file and ran md5sum on both, the source file and destination file. To my surprise, even though the file size was different, they had the same md5sum. Now, let it be known that the source machine is a production server and the dir i rsynced was being used, serving pages to the web. I googled about this and came up with stuff like open descriptors and holes. I don't understand this stuff and was wondering if this was really the case. What are those if it is the case? And my backup copy is 100% identical right? There are thousands of files and I ran md5sum only on couple. Can I take comfort that when time comes, I can restore using my backup without any problems?
View 3 Replies
View Related
Apr 17, 2011
Can anyone tell me that how to get information about stack, allocated by kernel to a running process? for this ,is there any api function,any system call is available in ubuntu 8.04 ?
View 2 Replies
View Related
May 22, 2010
Both gThumb and Geeqie image viewers have limited maximum size for thumbernails. Is there any way to work with bigger thumbernails (or slide photos)?
View 3 Replies
View Related
Jul 28, 2010
As I have a opt/sybase folder in Linxs I want to increase the size of this folder how can I increase it? For example the size we mention at the time of installation Sybase is 2gb now I want to increase by 4gb how can I increase it.
View 2 Replies
View Related
Apr 6, 2010
I installed Kubuntu 9.1 and then I downloaded Xampp of Linux to the desktop, however I have to extract it to /Opt/lampp. And that's the problem. I don't seem to have access or permissions to ?? I tried right-click | properties | permissions but the fields are greyed-out to make any changes.I am logged-in using the same user id/password I created when installing K9.1.
View 14 Replies
View Related