General :: Find Continuously Growing Files In The File System?
Apr 23, 2010How to find continuously growing files in the file system?
View 11 RepliesHow to find continuously growing files in the file system?
View 11 RepliesI've already asked this in the mythbuntu and didn't get an answer there so I'm trying here. OK just added another 2Tb WD drive to my mdadm controled RAID5 array, and the reshape is finished:-
Code:
Code:
Mar 28 18:28:36 alpha2 kernel: [104357.343421] md: md1: reshape done.
Mar 28 18:28:36 alpha2 kernel: [104357.525114] RAID5 conf printout:
Mar 28 18:28:36 alpha2 kernel: [104357.525119] --- rd:4 wd:4
Mar 28 18:28:36 alpha2 kernel: [104357.525122] disk 0, o:1, dev:sda2
[code]...
I made a BIG mistake and tried to print a PDF document from Evince 2.28.2. Evince sends a garbage filled print job to the queue that results in nothing but garbage being printed. I noted the problem an tried to kill the print job but could not because it does not show up in the print queue. It appears that the printer sent an error back to the print system and the print system now continues to print the job forever - that is, anytime openSUSE is running, even after reboots.
Note that printing the test page works and that the printer also works fine with gedit and OpenOffice.
If I go to Control Center -> Printing then right click on the icon for the printer and click View Print Queue, the print queue is empty.
In YaST Control Center -> Printer -> Policies the "Specify error policy" is set to "abort and delete the job and proceed with the next job." The "Apply this error policy to all local printer configurations" checkbox was checked when this policy was set. If this is what should have ultimately killed the job it did not work.
how to stop the print system from trying to print this garbaged print job. It has been trying to print the same garbaged print job for the last five days!
I am new to Linux. I am using tar to backup my emails to a server. I would like to automate this process to routinely backup my emails periodicly, however, i keep running into a problem: I start in the dir I would like to create the tar file (dir size = 240MB). I enter the following command
tar cf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages"
file size = 234MB
When I would like to backup my emails into the previously created tar file I use the following command:
tar uf bup.mail.llc.tar "/Users/d/Library/Mail/INBOX.mbox/Messages"file size = 462MB The backup command works, except the size of the original tar file grows, around twice the size. When I extract the updated tar file (file size = 462MB), the unarchived file is 240MB the same size as the original directory.
Why does the size of the tar keep growing each time i perform 'tar uf'?? I don't understand this
An external service I dont manage pushes mediafiles into a shared directory on my server. I need to move these files into their correct directories automatically. The problem is that if I run my script as a cronjob once every 3 minutes, I notice that the script copies files which are still on their way into my server. So I need to figure out how to have the script check that the files are complete (done downloading) before the script moves the files. This is what I got so far:
Code:
#!/bin/sh
# Script by proximity 280709.
# Locate correct directory
[code]....
Scenario: An IDE is set up on a Linux desktop box, editing PHP files locally. Every time I save a file, I want this change to appear on the linux server where Apache is running. The server has ssh (and samba and nfs for that matter).As a reference, when I edited files on Windows, I finally came over WinSCP as the exact tool I needed - WinSCP have just this feature present, with initial synch and then continuous update, using the filesystem watch service: "Keep Remote Directory up to Date".
On Linux, one could argue that sshfs could be employed to sidestep the need for synchronization entirely. On windows, a samba-share would do the same. However, I want the IDE to work with local files (on a SSD disk!), not having to go over the network to do PHP indexing and whatnots, which takes ages.But sshfs might be a part of the solution nevertheless - so that the continuous synchronization just needed to be done between two local directories.
Actuaaly i am creating watch on one directory in which files are continuously coming.Is there any command which can give listing of all files who have come in last 24 hrs.
View 8 Replies View RelatedUsed following command to find out the top 10 big files in the system But it is having its own limitations as it consider files and directories both.
Code:
I would like to get the following information.
1)top 10 big files.
2)top 10 big directories.
File size with human readable output.
As executing
Code:
But when i add -h option for human readable file size its giving me wrong output.
Code:
How to find file by content on linux by command. Example: i want find file contain word "helo" on my computer (OS: LINUX)
View 4 Replies View RelatedI work for a company that makes portable devices running Linux and I was recently asked to make the underlying file system read-only for "security" purposes. Since the distribution is based on LinuxFromScratch, I know that very little writing happens at run time. So, even if the device runs on a usb flash device, I doubt that putting the root file system RO will be that beneficial. I am actually more concerned about a process actually breaking because it cannot open a file in RW mode than a process going rogue and filling the root file system with log files, etc. I'd really like to ear what kind of advantages disadvantages there really is with read-only file-systems.
View 6 Replies View RelatedI use find / -name myfile to search files. But it will print out a very long list such as follow:
Code:
...
find: /var/empty/sshd: Permission denied
find: /etc/audit: Permission denied
find: /etc/httpd/conf/ssl.crl: Permission denied
find: /etc/httpd/conf/ssl.crt: Permission denied
[code]....
We have recently built some RAC (OS:RHEL55) servers and after the Oracle guys have installed their application, somehow the directory / is using the maximum space. I contacted the Oracle team & they say that their RAC installation doesn't create any files in the / directory. This is the o/p of '/' directory file system:
Filesystem Size Used Avail Use% Mounted on
/dev/mapper/datavg-vol2 498M 382M 62M 88% /
Also, when I checked the file sizes, I found that the following files were taking more space:
/etc/selinux/targeted/modules/active/
-rw------ 1 root root 17M Nov1 base.linked
[code].....
I don't know what these files are doing there, when I did a cat and cheked, I found the files containing this data:
nf_tre--
stem_dbusd_var_run_t...and some stuff like this Unable to decide whether or not to remove these files. Also, is there any way to find out what files are taking more space and whether they can be deleted or not? in order to free up some space in the / direcoty. As there are 10 RACs that we've build, I got to do something to fix this for all of the 10 servers.
Found the below from RedHat Knowledgebase
The Completely Fair Queuing (cfq) scheduler in Red Hat Enterprise Linux 5appears to have worse I/O read performance than in version 4. It appears as though the Completely Fair Queuing I/O scheduler (cfq) has a regression and thus exhibits reduced read-side throughput which can affect performance for both local and NFS mounted file systems.
One way to mitigate this is to set the cfq's slice_idle parameter to zero. To change this value, execute the following command echo 0 > slice_idle in the /sys/block directory appropriate for your situation, as shown below:
echo 0 > /sys/block/hda/queue/iosched/slice_idle
We are using NFS file systems in RHEL 5.3. I would like to know how to find which /dev/Device is being used by the NFS file systems, so that I could try setting the slice_idle to '0' to see if there is any difference in performance? In /etc/fstab I only see the actual NAS volumes for the NFS file systems.
I have this little problem. I wanna backup a big file that is constantly growing. Is there any way to make one backup and then have some way to take incremental backup of it?
View 2 Replies View RelatedIs there some file in Linux that enumerates and describes mount options for file systems like /etc/services describes ports?
View 2 Replies View RelatedI need commands which five me the following details abt all the file systems mounted the linux box
type
mount pt
file system id
I'm new to this and I'm sure I've just gone and done something silly. I used pendrivelinux.com's live usb creator to install ubantu 11.04 on my USB. I tried booting my laptop with this, and i get an error message saying "unable to find a medium containing live file system" and I cant get any further than that. The laptop I am using is a Dell XPS with an i7 processor and 6 gigs of memory.
View 6 Replies View RelatedCurrently running 10.04 and Cacti latest version and although my RAM is only just over 3/4 used my swap file is growing. If I leave the box running for about a week the swap fills and the system grinds to a halt.Is there some way of seeing what's in the swap so I can debug the problem and get my system more stable again?
View 5 Replies View RelatedI need to delete all *.trc files that are older than 30 days and I am getting a "Argument list too long" error. There are other files that should not be deleted which is why I am using the "*.trc" and newer files need to be kept as well. I have seen other postings but they do not cover both of the conditions. Below are 2 of the many attempts at doing this but I cannot get this to work.
find *.trc -mtime +31 -print| xargs rm -f {}
find *.trc -mtime +31 -print -exec rm -f {} ;
I had been trying to install Ubuntu 9.10 after formatting my machine. When i select install of the welcome screen, the logo blinks for few minutes and then displays a error as
Code:
Unable to find medium with live File System
Then goes to busy box.
"The same thing appears with Ubuntu 8.10, which i was using before!"
With what software I can slow down my HDD. With Power Manager if i change from Performance to Aggressive Powersave my HDD still working all the time and it's hot all the time on my laptop.
View 1 Replies View RelatedI'm having problem copying files from my laptop (windows xp) to moxa (linux) using ftp.
The error im getting is "533 read only file system".
Tried reboot and fsck but problem remains.
mount output below:
i am afraid that some one palying in my fedora 12 system , " hacked " first i got this message when i log to my broadband internet provider
Quote: You have asked Firefox to connect securely to ******, but we can't confirm that your connection is secure. Normally, when you try to connect securely, sites will present trusted identification to prove that you are going to the right place. However, this site's identity can't be verified What Should I Do?
If you usually connect to this site without problems, this error could mean that someone is trying to impersonate the site, and you shouldn't continue.
[Code]...
then when i log to yahoo mail , my system hang for a while then continue. how do i know that my system is secure ? what are the steps i should follow to secure my system ? i am still newbie and i dont have time to spend it in search in google or reading blogs
My main hard disk died and I replaced it. After installing windows in a small partition in /dev/sda, I thought I will try linux mint and went for it. (I need windows to play AOE, but ubuntu is my primary OS)I didnt see the options properly or some distraction, I choose the "install alongside windows" option probably expecting it to install it in the unallocated partition next to the windows installation. I had completely forgotten my second internal drive /dev/sdb which has the backup data. Linux mint went and installed itself on that drive.
Is there a way to recover individual files from the second harddrive. Now if I boot or open it through live cd, all I see in the linux mint file systems. I want to aleast recover my CV/resume from the second drive. The second drive is a single ext4 file system The old drive is completely dead, it doesnt even get recognized when I attach it to SATA.
There were some files residing on my ext3 file system, using Ubuntu as my linux distribution. Yesterday I formatted the hard drive using a windows install CD, rewriting it with a new NTFS partition. I'm willing to restore my personal files deleted due to this format.
View 3 Replies View RelatedI have searched many forums but did not find a way to FIND A CORE FILE WITH OUT USING FIND COMMAND.
View 11 Replies View RelatedIs there any way to find the core files with out using the FIND command?
View 1 Replies View Relatedshell scripting in Fedora14I want a script"Find in curent folder for files, and it copy first file he find with name gived by user, if name already exist then echo error message and finish"command usage " bash scriptname copyASname"
smthing like Code: #!/bin/bash
for files in /home/user/*
do
[code]....
Code:
if ! [ -f ${PATH}/myfile.txt ];then
echo $(date +%Y%m%d"_"%H%M%S)": Nu am gasit fisierul ${PATH}/myfile.txt"
ps -fxu pin | grep "/usr/local/coreutils/bin/tail -f ${LOG_PATH}/x.log$" | awk '{system("kill "$2)}'
cat ${LOG_PATH}/x.log | sed -n -e '/LONG/{x;1!p;g;;}' -e h > ${PATH}/myfile.txt
[code]....
I have a continuously growing log file (x.log) in which i have to look for certain lines that contain "Long". The line above each line containing the word "Long" it contains a time stamp. I want to extract each line containing the time stamp into myfile.txt and check the difference between time stamps. Whenever there is a difference i need to run another script (${CAL_PATH}/${APP_NAME}), then sleep 1, then continue searching. Lines with "Long" do not appear continuously, but in blasts. The script runs fine until the first pause encountered. Starting with the first pause, tail -f doesn't write in myfile.txt anymore. Can someone help me understand why "tail -f" it stops writing into myfile.txt? Or does someone know an alternative to "tail -f" of achieving the initial scope of the script?
I don't know what I did differently before this happens. Now I cannot access files or see icons on the desktop. The "starting file manager" continuously loops in the bottom panel, continuing to load forever. in the system monitor, if I kill the process "gnome panel", it resets itself, then continues the process. Only thing different I did was install that one thing that gives you System->Preferences->Windows by hitting "yum install control-center-extra".
View 3 Replies View Related