I want to list my folders and subfolders (recursive) and also show the size of the files in terminal. I started using this:
Code:
ls -h -R > /test.txt
I got everything but not the size of the folders.
Then I tried this:
Code:
du -h --max-depth=1 > test.txt
Suppose to show me everything, but I can't see subfolders. And this command do not accept recursive. How can I show the size of the files and folders like the second command, but including the subfolders?
In Linux bash shell, for a given directory, how can I list:The create date for that directory The number of files in that directory The number of subdirectories in that directory.
Let's say I want to get the size of each folder of a linux file system. When I use ls -la I don't really get the summarized size of the folders.If I use df I get the size of each mounted file system but that also doesn't help me. And with du I get the size of each subfolder and the summary of the whole file system.But I want to have only the summarized size of each folder within the ROOT folder of the file system. Is there any command to achiev that?
I want to get the size of each folder of a linux file system. When I use ls -la I don't really get the summarized size of the folders. If I use df I get the size of each mounted file system but that also doesn't help me. And with du I get the size of each subfolder and the summary of the whole file system. But I want to have only the summarized size of each folder within the ROOT folder of the file system. Is there any command to achieve that?
I am using secure delete to remove files from a Debian Linux PC. However, secure delete does not remove folders. This has lead me to look at writing a script that would move files to a predetermined folder for deletion. My plan is as follows:I have a folder on my desktop called shredder where I move the contents of the waste bin to. The script needs to identify all files within the folders and sub folders, within the shredder folder, and move each file to the shredder folder and then delete the folder. At this point secure delete can be used with a command like shred -v -u *.*on the shredder folder.The problem I have is in creating the code to move files from the different folders and then deleting the folders. Note that the names of the files, folders and subfolders will not always be known
Is it possible to make a script that makes a tree of all folders and subfolders and outputs it to either a .txt file or .pdf? All folders except from one shall list 2 levels. The except folder all the way.
lets say I have a project that have generated lots of xml files. Though all these xml files point to a location with the text name TEXT15. I want to change all the files that containts TEXT15 and change it to TEXT16. This actually works for files in a folder but not recursively in all the entire files....perl -pi -c 's/TEXT15/TEXT16/g' ./* but I have many subfolders and within this more subsub folders....i just want to do this recursively.
I have a 1TB external hard drive. I would like to create in it 10 folders:
Code:
I would then like to permanently mount each folder to its machine (I have 10 machines connected through a switch, so each machine will have a folder that is mounted to ONE of the 10 folders in the external hard drive).
My questions: (1) Is this a good configuration? are there better ideas to give individual machines more space without replacing their hard drive? (2) How do I limit each one of the folders ('folder1', 'folder2', ...., 'folder10') to a size of 100 [GB]? I don't want one folder (say, 'folder1') to grow in size and 'steal' the space designated to the other folders.
if there is an application available to generate a list with files and folders from a location, like a hard drive or a folder? The list could be in any format, even a text file would be just fine.
I don't know if this is a known issue, but when I want to copy/move a file to a folder containing lots of sub-folder displayed as list, I can't do a simple drag&drop, because there is no empty space to move this file. It will always go into one of the sub-folders. So obviously I have to go up one folder or to change view from list to Icon or Compact. Is it a known issue? Is there a solution? (Using GNOME nautilus 2.30.1)
I'm successfully connecting to my Vista desktop via my laptop running MM;I can view my shares but not all my data. All top level shares seem to be visible, it's when I go deeper that I notice not everything is showing up. The folders are accessible if I type the full path, and then again I have the same issue once I'm inside that folder where not all data is visible.
In Nautilus setting: Edit>Pref>Preview Tab.I set preview for all files under 4GB which is the upper limit I can set.This made more files visible but not everything.I attempted to divide the directories in order to halve the size of the folders, this helped but not completely. The directories contain movie and picture files. The picture files are easier to divide more equitably between folders, but the movies being so large, managing folders sizes in that way doesn't help. I really don't want to divide my files according to this problem anyway. Sorting does not affect which folders and files are shown. Nothing is hidden, options to show hidden files on amd off, no help.
I want to know how much damage a user can do on my system if he decides to delete everything (or write to in case of corruption).What command or script might i use to check this?
Is there any Linux application for finding the folders with the most number of files? baobab sorts folders by their total size, I'm looking for a tool that lists folders by the total number of files in it.
The reason I'm looking is because copying tens of thousands of small files is excruciatingly slow (much slower than copying a few large files of the same size), so I want to archive or delete those folders with high file counts that that will be slowing down the copying (it won't speed things up now, but it would be faster when I need to move/copy it again in the future).
Just wondering if it's possible to change the size of icons when Nautilus is in List View. I tried playing around with gtk-icon-sizes in the gtkrc file of my theme, but there doesn't seem to be a setting for it there. I could just zoom in/out, but when I have the icons the size I want them to be, the text is too small (see attachment).
I want to know how to get the details of all files in my system more than 20mb or some specific amount of file size in Ubuntu terminal by using commands.
I am in the process of customizing my desktop, and I have run into a problem. As a programmer, I will invariable end up with tons of windows open. I would like to expand the gnome-panel's Window List's y size to allow more applications to show without being crunched together.
That's the way the desktop currently looks, notice the tiny Window List in the top left. That's as big as it gets. I want to expand it all along the entire left side. I have a feeling the answer is in gconf-editor, but i just can't find the right field to modify.
I thought there exists a built-in function that gives total size of a list object in byte. So using the function produces 1 for a1, 3 for a2. I haven't found such function or module yet.
I want to generate a temporary random list from a directory of files and then determine the size of an arbitrary block of files from this list (say 1-25 or 26-50) and add their names to a file along with some other info for each name. I can generate a random list with file sizes like this: ls -l | sort -R | cut -d " " -f 6 but i'm not sure how to add up the sizes of just a certain block of these files and at the same time save the file names.
I am looking for a utility that would do the following:1. Be run manually on a list of files whose sizes should not change, to get a control file containing the sizes of each file.2. Subsequent manual runs would report any changes in size of any of the files in the list, and allow option to accept the new sizes.3. Be run as a cron job to check for changes in the file sizes and send an email alert if a change has occurred since the last time it was run.The purpose is to detect possible hacks of key files on a website. It would not include files expected to change, but just those that should not change. It would be run manually a few times to get the control list one wants to monitor.
I was recording a DVD when I wanted to look at a folder to see if there was a file there. I went to "places", as I always do, and a window poped up asking for a password. This has never happened before. I entered the password and went to the folder. I always view folders as "list", then they displayed as "icons". I finished the recording and re-started. It still appears the same, and when I open programs I notice little differences in the way things look and feel. The reason I was recording DVD's was to get some of the movies off my hard drive. It was down to 81Gbs free, It now says there is 365 Gbs free. I have searched everything and I see nothing missing.
i have a folder with 250 subfolders and each one of them has (at least) one image in it (along with other stuff)how can i 1)copy all the images from those subfolders and paste them into one folder together (other than by hand, obviously)?2)optional : copy only the images of a certain size and above?