Is there any Linux application for finding the folders with the most number of files? baobab sorts folders by their total size, I'm looking for a tool that lists folders by the total number of files in it.
The reason I'm looking is because copying tens of thousands of small files is excruciatingly slow (much slower than copying a few large files of the same size), so I want to archive or delete those folders with high file counts that that will be slowing down the copying (it won't speed things up now, but it would be faster when I need to move/copy it again in the future).
I just had a bad experience with a server, and now i have a 17GB lost+found. It appears a lot of the stuff in there are folders which are empty. Since those really serve no purpose as far as recovering data from what I can tell, is there a slick way to delete just the empty folders from my /data/lost+found folder, leaving me with just the stuff I truly need to look through?
I have a USB drive that has a TON of folders on its root level. I want to remove all those folders and their contents except three of them. I know if I do rm -rf that will kill everything, is there a way to exclude three folders, say folder1 folder2 newfolder, and do it all in one statement?
How would i delete all the folders access one i specify.for example if a there are 4 folders namely a b c and d. How would i prevent deleting folder b .Whatever folders other than b should be deleted?
I have been forbidden to enable automatic updates on our Ubuntu servers, for both security and regular packages.When I log into any of my four Ubuntu servers, the welcome message contains this:
39 packages can be updated. 26 updates are security updates.
However, when I run the Nagios plugin that monitors APT, I get:
% /usr/lib/nagios/plugins/check_apt APT WARNING: 33 packages available for upgrade (0 critical updates).
I need to know how to properly detect that there are pending security updates, and regular updates. Once I can do that, I plan to write a Nagios script that will return WARNING for pending regular updates, and CRITICAL for pending security updates.
I am happy to report I successfully setup a cloning station using Ubuntu 10.4 Lucid Lynx and Clonezilla. I learned the hard way you have to restore images to a hard drive of a larger size than the drive the image was ripped from. I had already ripped 6 or more images and have them in the Home_Partimag folder and cannot delete or rename these images. I want to remake these images and rip them from smaller hard drives. Any commands I can use to be able to do this?
I want to know how much damage a user can do on my system if he decides to delete everything (or write to in case of corruption).What command or script might i use to check this?
i have been copying dvd's and now i do not have enough room in my temporary folder,but i cannot find out how to delete the folders and make space to continue
I am administrating a lab in a university and every semester we need to delete all the home folders of the accounts for the next semester. I would like to make a bash script that does this automatically and having trouble with it. Note that I am writing my very first bash script. What I need to do is make a script to delete the following:
Delete everything in /home/$exp$num/$dir when "exp" could be either "rt", "ic" or "sp". "num" could run from 1(single digit) to 45 and dir is "profile" and "work".
This is what I tried to write:
Code:
#!/bin/sh cd /home for exp in "rt ic sp" do
[code]....
What seems to be the problem is the reading of "$exp$num" as a joint expression.
I have an archive directory that needs to be cleaned up once per quarter. The top level (/data/archive/*) directory names change daily, as well as the subdirectories and the filenames (the application names everything according to date). Also, there are two top level directories, bin and incoming, that we can't touch. I want to write a shell script that loops through the 15 or 20 top level directories and deletes all files and subdirectories older than 3 days (skipping the bin and incoming folders). Can someone get me started on a script? I am kinda new to shell scripting.
In Linux bash shell, for a given directory, how can I list:The create date for that directory The number of files in that directory The number of subdirectories in that directory.
Many folders within a subdirectory some of which have lots of data in and some of which have only one specific file called produkt.fil inside.I need a command to find and delete those folders that contain ONLY the file produkt.fil - if other files exist (doesnt matter what they are) then they should be left alone. Note: produkt.fil exists in all of the folders always.
I don't know how to explain this but the free space indicator in folders shows a different number than in Gparted. The Picture pretty much shows it all.
Although I've been using Ubuntu for a while, I haven't really done any scripting thus far,but now find myself needing to do (what I thought was) some fairly simple stuff.However, despite much searching, I have been unable to find a solution to my problem.I need a script which will delete all of the files/folders from a directory except those I specify. From the command line, I am able to use:rm -r !(constant|system|*.msh|)
which works fine, deleting everything except the 'constant' and 'system' folders (and everything in them) and any files ending in '.msh'. I have tried to put this into a script file, specifying to use the bash shell, and escaping the required characters, but I just get the message 'cannot find !(constant|*.msh)'.I have read that using the find command may be a better way of doing it, but so far I have not been able to make the command work in the way which I need (I have easily been able to delete ALL the files/folders in the directory!).
In my project i cannot determine the number of check list initially. I will know dynamically during execution.How to specify the number of check list dynamically in zenity.
I have an FTP server on my network (currently ran on IIS). When I connect with Nautilus to FTP and try to delete folder (files delete works) I get error.I tried Filezilla and Gnome-Commander as clients - they cat delete directories just fine. I tried to solve this problem for months because I like the convenience of having my FTP bookmarked in Nautilus and use it from there.Today I finally found a reason for Nautilus not being able deletedirectories, but I can't find a solution to that. I've checked FTP logs and saw that all working ftp-clients are performing RMD on directory name only, while Nautilus always prepends directory name by / symbol. So, let's say I'm trying to delete folder call "Test"Filezilla and Gnome-Commander are logged as doing RMD on "Test", but Nautilus was logged as doing RMD on "/Test" and server replied that "/Test" folder doesn't exist
Currently I am doing a java project and I renamed some file/folders to start with '.'. i.e. ".project", ".classpath", ".settings/"
I can see them on my desktop but i cannot delete them. In the terminal, "ls" command does not show these files exist. "rm" will also say cannot find these files. How do I get rid of these files/folders?
I have a Ubuntu file server with a mix of 30+ users ( mix of windows and linux ).All are members of the same group. All need read write create access. I want to prevent deletion of certain key folders. How can I achieve this ? sudo chmod -R nnnn ??
I have installed a new system on my desktop computer, and would now like to delete files/folders from the web interface, as they no longer reflect the files and folders on my desktop PC. I use my desktop PC to sync files and folders one way to my laptops.