i need help in this issue how to find files with unusual size and with unusual names of EX : just dots, names ending with space(s),names containing shell wildcard characters, names containing non-ASCII (control) characters
I made an account under freeshell.org and it has been very satisfactory so far. I recommend everyone getting an account under freeshell.org. But anyways, how do I find files over, for example, 500 KB, in the entire, my shell account?
am trying to write a shell script to find the size of a particular log file and if the log size grows, script should mail the changes to the administrator or a any user so script should monitor the log file continuously in a time interval, how can i do that?
I tried with these codes to find the file size but it throws me error says command not found
How to find the filesize of a flash which is not exactly stored in the temp but redirected by a fd. flash31 -> /tmp/FlashXXvsg1uY (deleted) directory is /proc/processid Chrome is downloading a flash and I can see in the proc directory the flash file, How can I see the current file size of such a file ?
I created a VM disk image with kvm-img, but I forget what was the max size of that disk image when I created it. Currently, its size is 6.2G, I want to install some large packages in that VM, so I want to make sure the disk image can expand to an adequate size.
I am trying to do a find/grep/wc command to find matching files, print the filename and then the word count of a specific pattern per file. Here is my best (non-working) attempt so far:
Is there a way to specify to find that I only want text files (and not binary files)? Grep has an option to exclude binary files, so I thought find probably has a similar feature, but I've been unable to find it.
I am having a few situations to which I do not see any thing in du man pages.Quote:1) I want to see files in a sub directory which are larger than a particular size only.2) I use du -sh > du_output.txt I see the output as described for option -s and -h how ever what I am more interested is if the output comes in a format which is say for example
I know how to search for normal files but can you let me know " How to search for 5 setuid files on the system. Also explain, for each file, why setuid mechanism is necessary for the command to function properly"
I have a Canoscan LiDE 30 scanner which works fine in Linux. I haven't used it in months, and never since moving from Jaunty x86_64 to Fedora 11 x86_64. This evening I suddenly need to scan two pages to send to someone via e-mail, and the recipient is desperate to receive them. Naturally, Xsane has decided this is the moment to screw me.
The problem: I cannot figure out how to tell Xsane that the page size is 8.5 11. It insists on scanning only the left side, 3.99 x 11. I cannot find a page size anywhere in any of the buttons.
In Preview I once got it to scan the entire page, but could not figure out how to save the image to file. Subsequently I have not been able to figure out how I did it. In any event, scanning from the small main window always scans 3.99 x 11. Adjusting the scan area manually in the Preview window allows me to change the size area in the small main window to 8.5 x 11, but when I then scan from the small main window it reverts to 3.99 x 11.
I have scanned the document a hundred times at least, each time trying different buttons, but nothing has succeeded in getting the full page.
i manage to delete some files from the system. now i need to recover them.. i know the inode # (through ext3undel) and also the size.Quote:Unfortunately, we cannot automatically obtain the name of a deleted filefrom Unix file systems - since the connection between the iNode (whichholds the MetaData, including the file namee real data is droppedon deletion. However, we can obtain a list of names from the deleted files.How can i use this information to recover the files?Also can i search the text from a partition? (file don't exists). As i need figures
We are currently running Redhat 5.4 64Bit (Build 2.6.18-164.e15) however having issues with tar files that are greater than 2TB in size.
We have been told that upgrading the kernel to support the ext4 file system (supported in version 5.5?) and mounting the current 10TB nas share as an ext 4 file system may solve our over 2TB file size issues. Are their limitations within the ext3 file system that cause issues to files greater than 2TB and if this is the case Do we have to update the kernel (complete rebuild), or can we load a package on the redhat box to support the ext 4 file system Where do we get the upgrade/package. I have logged onto the redhat site and can not find kernel updates but can find the full installation packages. We are not connected directly to the internet and do not have access to the update repository
I want to generate a temporary random list from a directory of files and then determine the size of an arbitrary block of files from this list (say 1-25 or 26-50) and add their names to a file along with some other info for each name. I can generate a random list with file sizes like this: ls -l | sort -R | cut -d " " -f 6 but i'm not sure how to add up the sizes of just a certain block of these files and at the same time save the file names.
I am newer to Linux ( using Ubuntu 10.04) : I have noticed that during replacement of a file , no date and size of the new and old files are shown in the dialogue box so how to show that ( like the one in windows)
I know that it is easy question , but i really don't know how to do that , by the way I have checked folder preferences and system --> preferences but i did not find something for that
I am using wget (httrack can also do that ...etc) to downloading entire website , but is there any way to find out the size of the website before starting the download.
I have a lot of mp3 files without any ID3 tag information. This is very annoying, because my iPod does not show them correctly. My Banshee shows them as "unknown artist" and the title. I would like to find them all in a bunch to batch update them.
We have a huge amount of duplicate files in a folder and I would like some pointers on to writing a bash script to create a list of the duplicate files. I've seen examples that check for the md5 sum of files... but I dont need that, the file name is enough.
following are log files which are existing, I have to delete one day older files from this location but when use above mentioned command it won't print one day older files, as i understand "-mtime" modified time, "+1" means one day older. am i correct?
Code:
-rw-rw-rw- 1 ablddb dba 268435456 May 30 17:11 mtsDB.log126985 -rw-rw-rw- 1 ablddb dba 268435456 May 30 17:17 mtsDB.log126986 -rw-rw-rw- 1 ablddb dba 268435456 May 30 17:23 mtsDB.log126987
I'm trying to find all java files in bash that contains the method "assign()".I would like to retrieve the same list except without the Test* files. How can I do that?
In Linux bash shell, for a given directory, how can I list:The create date for that directory The number of files in that directory The number of subdirectories in that directory.