General :: Find The No Of Files Cached At The Time Of Open & Loading Folder?
Feb 5, 2010
when we do enter on a folder it take some time for loading the folder depending on the no of entries in the folder . If the folder has more entries it take more time to load and if less no of entries then correspondingly less time . the delay in loading the folder varies due to reading of the folder entries in advance . SO what i want to know is that what is the MAX no of entries read in advance while opening a folder in linux and also how can we calculate this
So everytime i try and use the "yum" command. for some reason it doesn't go past this point: "loading mirror speeds from cached host files". i cleaned up the cache and rebuilt the OS again, and im still getting this problem. would this be a problem with my internet connection? I'm using CentOS 5.5.
I've a folder named "lynx" in my desktop . It contains ebooks - pdfs,chms etc . The problem is sometimes when i open the folder in nautilus , the folder keep loading forever and sometimes it shows two or three files and keep on loading . But at some other times , folder contents are displayed just as any other folder(i.e normal ).
I was created one folder in linux with current time was 1978(For example). I was moved this folder to usb(FAT32 file format).While seeing this folder in window its not showing the folder time created time stamp, because the USB file system only support the year after 1980 . But again i am putting the same folder in linux ,its showing the correct time stamp.How is it possible? Because FAT32 only supports timestamp after 1980, but still its showing 1978 in linux system
We have a script that FTP files 3 times a day, once ant 02:30, 04:00 and 13:00. Once the process runs it puts a copy of the file sent in the processed folder. What I'm trying to do is check to see if the files are there and if not send an alert /email. The file names are IVF_20100806_*.150, PLAZ_ 20100806_*.151, TRAN_20100806_*.152 and TRAN_20100806_*.151
when i used windows there was this wonderful editor named Notepad++.it was perfect(it still is) some of its best and useful features of it (for me) was:
1-open all files in a folder when drag and drop the folder on it 2-search and replace a statement in all open files 3-have an extended mode which include special characters like
and so on.. i want to know if there is an editor with this feature in ubuntu?
I cant open xls files in a shared folder using open office 3.1. Its all started when i install a new open office version. when im using open office 3.0.8 i can open all the files but now i cant even open it. Its always display untitled documents.
I'm trying to find a proper command to move a certain set of files according to date/time range. I am thinking that the command should be something like:
I tend to watch the odd movie, and always use vlc. As such, one time I checked the box so that vlc would open by default. now however each time I go to open a folder, vlc starts up! How can I undo this?
I have been using ubuntu for a while and i like it a lot, im a web developer and i have windows xp installed in virtual box, i moved completely to linux and just use windows to test in ie, it had been a while since i didnt use windows and i had to use in the last few days and noticed how much faster it is, the thing that bothered me the most is when opening folders in the desktop or the recycle bin, in windows its instant, in ubuntu opening a folder takes a long time to open nautilus, is this normal or is my installation bad, any comments are appreciated, i dont want to abandon ubuntu, i really like it but it really bothers me that nautilus is so slow to open.
shell scripting in Fedora14I want a script"Find in curent folder for files, and it copy first file he find with name gived by user, if name already exist then echo error message and finish"command usage " bash scriptname copyASname"
smthing like Code: #!/bin/bash for files in /home/user/* do
I ask question about my iPhone 3GS here because it run in linux too.I think it's Debian.it if u happen to know the answer. I want to keep cache downloaded from Cydia for reinstalling later.
I am trying to find a nightly backup if it was successfully copied over, rename it and curl, but it's always passing the check even if the file is older than specified. From the command line it does as it should. Example is here;
Code: find /backup -type f -mmin +4440 -exec echo "found" {} ; - nothing returned (good). Then I change the time
I've using RedHat/Fedora for years now, and every now and then I encounter the following situation :
I open a folder and it's empty. The folder was containing files and I'm 100% sure I didn't deleted them myself. Each time the folder is deep inside the hierarchy and is among other untouched folders. Sometimes it's a folder I never use, sometimes it's a folder I use almost everyday. The missing content is not large (a few regular files).
I'm currently running F13 but I've seen this behavior before on previous versions. This is kind of scary all my work is there and my backups are also done on a a linux backup server.
I'm puzzled, I cannot see any specificities to these folders, I had no crash or cold reboot, nothing I see can explain that. Could it be related to ext3?
what i wanted to do was find all the files with a specific name from a tree, sort them by modification time and have their directory appended to them so that i knew where they were (because they all have the same name). i tried a whole bunch of different things and finally did this:
this did the trick pretty well, but as you can see it is far from elegant and i think i'm doing some things wrong and kludgy
first thing i tried was "ls -lRt | grep world.sav" which worked except i couldnt distinguish the files because there were no directories. that took a lot of looking till i accepted i couldnt make ls print directories as well and append them to the files somehow that their relationship would be clear. i tried piping ls to find, doing it in reverse, passing them from grep etc. etc. until i read some more stuff online that got me using gawk and sort. the questions:
1. is there some other, more elegant and simple way to do this kind of detection and sorting?
2. is there any way to use a pipe after using exec? the semicolon seems to prevent this entirely, forcing me to use an intermediate file as above. i could just remove it later, but i'd prefer a straight piping.
With Windows XP, I just right clicked a folder/directory and pressed FIND then I could search for whatever file/folder name i wanted to. I could even do custom searches based on the size, modification date etc.How do you do this on Ubuntu? There doesnt seem to be a way to easily do it like this. So far i found PLACES -> SEARCH FOR FILES but that means I have to go into the directory i want. Where as I would much rather be browsing through directories and THEN want to quickly search in a particular folder. The SEARCH FOR FILES method in PLACES just wastes more time.
I was messing around with fed 12 yesterday (only on a test installation) and i've hit a snag. I installed openbox and tint2, nitrogen, obconf etc so i could have a #!-style session at startup. It all worked fine until i installed pcmanfm and removed nautilus. The problem is that i can't display files in my home directory, either using pcmanfm, a reinstalled nautilus, or in terminals. Every time i try to point a file brower there it just seems to get stuck searching forever, until i kill it. Weirdly in terminator i can do an 'ls' to see visible files, but 'ls -la' causes the problem again.
ps i thought permissions might have something to do with it, so i did a 'chmod -R 777' as root. it changed permissions for quite a lot of the files but then froze again, and now the problem persists.
In Linux bash shell, for a given directory, how can I list:The create date for that directory The number of files in that directory The number of subdirectories in that directory.
How would i go about copying all .jpg or .JPG files from a folder and all its subfolders to my /usr/name/pictures folder? I'm guessing I'd have to use some sort of .[jJ][pP][gG] to get all the pictures from other examples i've seen, but really not sure how to use that in a recursive cp.
I got a folder that I transfer stuff to all the time. the folder is in chmod 775 but when i upload folders and files, they are given chmod 700, but i want it chmod 775 everytime i upload something. so far i have logged in to my linux computer and did a chmod -R 775 to the folder every time i uploaded something to it. is there a function somewhere to make it 775 everytime i upload or can i have something run a script, so i don't have to go in and write it everytime i upload something?
I am trying to create a package for Moblin using the Moblin Package creator application in Ubuntu. The RPM package gets created successfully, but when I execute that package in Moblin OS in the terminal I get the following error message: Error while loading shared libraries: libgda-4.0.so.4: cannot open shared object file: No such file or directory.
I understand that the above libraries are missing on my system and I need to add them. My question here is do I have to add them manually after the installation of the package or can the package itself include these libraries and copy them to proper locations when the package is installed. In case the libraries can be included in the package I want to know how that can be done.
I have a weird performance issue with a centos 5 running a nfs server and a rh8 client. I think the fact that it is rh8 client should be downplayed. It is just that with rh8 client the performance degradation seems more clear. See test details below OS in server is Centos 5 x86_64 kernel 2.6.18-92.1.22.el5
1Gb connection between machines File to test over NFS is a 1GB file. First of all I wanted to measure how the network alone performs while using NFS. So in the server side I run a "cat" command on the 1GB file to /dev/null. Please note that the disk read speed is about 98MBs. At this point the file system has the 1GB file cached in memory. In the client side a "cat" on the same file gives me a speed of about 113MBs. It seems then that the bottleneck in this instance is the network and it is very close to nominal speed. So the network performance is really good. (BTW I know that the server got that file from cache because a vmstat or iostat shows no disk activity.)
The second test is reading from disk with no caching involve. In the server I flushed the 1GB file from the memory. For instance by reading another 5GB file and I repeat the same thing as above in the client (a cat on the 1GB file). Now, the server has to go to disk.(vmstat or iostat shows the disk activity). However the performance, now, is about 20MBs, I was expecting something closer so 90MBs. (since the reading speed in the server in the first test showed 98MBs).
This second test was repeated for ext2, ext3, xfs with no significant differences. A similar test using a RH8 NFS server and client gets me close to 60MBs for a 1GB file not cache by the file system in the serverSince network speeds and disk read speeds are not the bottlenecks ... what or where is the limiting factor then?