General :: Find Big Files On The Disk?
Oct 17, 2010How can I find big files on the my Linux disk with say more than 500MB in size?
View 2 RepliesHow can I find big files on the my Linux disk with say more than 500MB in size?
View 2 RepliesI am trying to do a find/grep/wc command to find matching files, print the filename and then the word count of a specific pattern per file. Here is my best (non-working) attempt so far:
wc `find . ( -name "*.as" -o -name "*.mxml" ) -exec grep -H HeightResizableList {}` ;
Is there a way to specify to find that I only want text files (and not binary files)? Grep has an option to exclude binary files, so I thought find probably has a similar feature, but I've been unable to find it.
View 2 Replies View RelatedI know how to search for normal files but can you let me know " How to search for 5 setuid files on the system. Also explain, for each file, why setuid mechanism is necessary for the command to function properly"
View 1 Replies View RelatedI know I can do find . -type f, but that includes binary file and I couldn't find a way to exclude them with find
View 4 Replies View Relatedi have installed ubuntu version 9 and i cant seem to find where hard disk partitions are,what do i do?also what do i do to install the webcam and to change from gnome to Kde enviroment!!i also installed virtualbox but i seem not to find the icon
View 3 Replies View RelatedHow do I check my hard disk Revolutions Per Minute (RPM) speed from a Linux shell prompt without opening my server case?
I referenced some other articles. they give only model number, serial number and disk space, but I need Hard disk RPM speed using shell script. any other Java program
I have written a script that triggers a mail if the server load average goes beyond a specific value.
The mail contains following field Current Load average. Top 10 CPU utilized processes.
Code:
ps -auxf | sort -nr -k 3 | head -10
Top 10 Memory Utilized processes.
Code:
ps -auxf | sort -nr -k 4 | head -10
But the problem is that whenever there is any disk related activity happens the load gets high and the command (for ex.mkfs.ext3 /dev/sdb1, dd,scp,cp)which are the main cause behind the load average doesnt get displayed in top 10 CPU/Memory Utilized processes.
Is there any way of finding top 10 processes for Disk related activity?
Is there any way to find the time required for accessing a block
from disk?
I want to find all users shown in the /home/ directory whose disk consumption is more than 500MB. The following command works as expected.
cd /home/ && du */ -hs
68K ajay/
902M john/
250M websites/
From the above example, only 902M john/ should be returned.
How can I make the find command output the same results?
I created a VM disk image with kvm-img, but I forget what was the max size of that disk image when I created it. Currently, its size is 6.2G, I want to install some large packages in that VM, so I want to make sure the disk image can expand to an adequate size.
View 1 Replies View RelatedI read this thread but
Code:
anisha@linux-uitj:~> su
Password:
[code]...
I am trying to get the total file size for certain files per directory.
I am using
find `pwd` /DirectoryPath -name '*.dta' -exec ls -l {} ; | awk '{ print $NF ": " $5 }' > /users/cergun/My Documents/dtafiles.txt
but this lists all the files in the directories.
I need the total per directory for all dta files.
I seem to be running low on disk space on my linux server. 'df' shows about 82% usage on a stock CentOS install with sendmail. I routinely delete old email directories, but for some reason, I stay pretty high in disk usage. Is there a fancy little bash script or something I could run that would find the biggest files and I could go get rid of them?
View 3 Replies View RelatedCode:
server:/home/mal # cd /backup/
server:/backup # du -hc *
17G export
4.0K lost+found
20M rman
8.0K sqlnet.log
4.0K test.sh
[Code]...
All my important files is already gone because I accidentally reformatted my 160GB hard disk. What software should I use to recover my files?
View 4 Replies View RelatedI am in the process of re-DL-ing the ISO as the checksums didn't match.
Mind you, that was with a shell extension in Win-Lose. Who knows.
Anyway:
I have the ISO (that I had previously) on a USB courtesy of UNetBootIn. All appears well until, quite quickly, I receive an error after the loading process which goes something like:
Code:
Error: cannot find disk at [hash-code - looks like 0ace5f etc etc, is about 12 char's in length.]Something very similar (but not same error I think) happened when I tried to do the same with my OpenSolaris or Fedora install. I.e., gets as far as the very beginning of the loader and then: bork.
At least one of them said "will reboot in 120 secs". Saves me the trouble!
How do I get Ubuntu's "Disk Usage Analyzer" to show me the hidden files?
It tells me my home dir uses 3GB, but only accounts for 525MB (the results of du -shc *). Can I get it to show me the other files that are using the space?
I use Linux. I coded a screenshot program some time ago and now I have 9 GIG of screenshots, 60000 JPEGs, most of them look pretty similar, and I have 300 MB of disk space remaining.
What are some good ways to start to compress batches of them (or all of them) in the background given the limited space? The problem with compressing the folder all at once is that I wouldn't have enough disk space for that. It seems the process needs to be broken down into chunks. So maybe something like: Get a list of all the files Add a chunk of the files (say, 20) to a compressed archive. Once it is done and saved successfully, delete the chunk of files
I removed my Ubuntu install and decided to replace it with Debian. I backed up the /home directory onto the Windoze installation on the other hard drive. That was a "home.disk" file. Now, I copied the file over to the Debian hard drive, and can't figure out how to recover the files. Is this possible to do in Debian?
View 14 Replies View RelatedMy main storage partition got full, so I'm deleting files to make room. However, df -k keeps reporting no space available on that partition (/disk). Here's the output of the command several minutes apart while another process is deleting a 30G of space:What can I do to make the space available immediately?
View 4 Replies View RelatedI'm running CentOS 5.5.
A couple of weeks ago, my 500 GB disk crashed after suffering an accelerating error rate for a couple of days.
Now that I have a new disk in place, I want to mount my old disk, which is (was) an LVM disk, as a second disk and recover files from it if possible.
The question is, how do I go about this?
If I mount the old disk:
# mount /dev/sdb1 /sdb1
the top-level directory shows only two subdirectories:
lost+found, which is empty; restored, which is an image of files recovered from old disk's predecessor.
'df' shows (or appears to show) that the disk is 11% full. Sorry, I don't recall how full the disk was before it crashed.
How do I get the system to recognize the LVM structure on the old disk & mount it?
Or is the directory structure too corrupted? Do I need to send the disk out to a recovery service?
P.S. Have aleady done some poking around with lvmdiskscan, which shows the old disk whether it's mounted or not, and vgchange -ay, which doesn't appear to do anything.
I got an USB hard disk, which is formatted in VFAT32 in window. I try to back up my file.tar to this hard disk. i got this error: writing `/media/PD03/xxx.tar': Input/output errorand the mount point is disappeared by itself.mount point: /dev/uba1the file I tar as a root... i copy the file as a root user as well.mount point info:/dev/uba1 on /media/disk type vfat (rw,noexec,nosuid,nodev,shortname=winnt,uid=500)
View 3 Replies View RelatedI have a lot of mp3 files without any ID3 tag information. This is very annoying, because my iPod does not show them correctly. My Banshee shows them as "unknown artist" and the title. I would like to find them all in a bunch to batch update them.
View 2 Replies View RelatedWe have a huge amount of duplicate files in a folder and I would like some pointers on to writing a bash script to create a list of the duplicate files. I've seen examples that check for the md5 sum of files... but I dont need that, the file name is enough.
View 4 Replies View RelatedSay someone logs into a Centos Linux box and SCPs a file from /root/ called 'textfile.dat'
Is there any way to log into the Centos Linux box, later, and see that textfile.dat was SCP'd from the box?
I used following command to sort one day older log files
Quote:
find /opt/TimesTen/tt_transaction_log/ -name "mtsDB.log*" -mtime +1 -print
following are log files which are existing, I have to delete one day older files from this location but when use above mentioned command it won't print one day older files, as i understand "-mtime" modified time, "+1" means one day older. am i correct?
Code:
-rw-rw-rw- 1 ablddb dba 268435456 May 30 17:11 mtsDB.log126985
-rw-rw-rw- 1 ablddb dba 268435456 May 30 17:17 mtsDB.log126986
-rw-rw-rw- 1 ablddb dba 268435456 May 30 17:23 mtsDB.log126987
[code].....
How can i print one day older logfiles?
To find the space occupied by files modified more than 4 years ago, i tried following.I am wondering if it is right ?
Code:
find /temp -type d ! -name ".*" -mtime +1460 | wc -l |du -sh
I tried this, but this sits there for long time (of couse the path i tried has lot of files) So i am not sure if this is right.
P.S.:
SHELL=bash
OS=RHEL5
find text within files? detailed at Code: man grep
View 2 Replies View RelatedI'm trying to find all java files in bash that contains the method "assign()".I would like to retrieve the same list except without the Test* files. How can I do that?
View 3 Replies View Related