Ubuntu :: 10.4 Making HUGE Log Files--running Out Of Disk Space?
Feb 2, 2011
I started getting errors about running out of disk space in root this morning. I hunted up what's taking all the space; var/log is 39GB (Ubuntu is installed on a 50G partition.) It's specific files that live in that directory, not subfolders. The files are:
My /var/ partition continues to fill up on all my servers, and it is because the logs in /var/log/apache2 or /var/log/mysql are being deleted during log rotate, but their file handles are being held open. Thus, a "du -sh /var/log" shows the correct values, but "df | grep /var" shows something much different.
It seems that the log files rotate, however if I run "lsof | grep deleted" it returns lots of files that are no longer visible in the directory, however refuse to clear themselves off the disk.
The only way I have found to make these log files go away (and thus clear up the disk space on the partition I should have) is to restart either apache or mysql, depending on which process has huge sized log files being held open.
Is it just me, or is this a big flaw in the way linux works, that it can't figure out how to release file handle for a log so the disk space can be reclaimed? This is happening to me a lot lately.
Here is some output from one of my web servers so you can see what I am seeing...
root@web49:~# df -h | grep /var$ Filesystem Size Used Avail Use% Mounted on /dev/sda8 9.2G 6.1G 2.7G 70% /var root@web49:~# du -sh /var
I am running the command on a Mac but due to it being a generic unix command and a command line query.. I thought I can write on this forum.. I am running the command
Code: df -h | grep '/dev/' I get Code: /dev/disk0s2 389Gi 62Gi 327Gi 16% / /dev/disk0s3 76Gi 24Gi 52Gi 32% /Volumes/Backup /dev/disk3s2 500Gi 47Gi 453Gi 10% /Volumes/Misc Note the huge space between the 1st and 2nd Column..
This is because currently I have some NAS drives mounted which are not showing due to grep. When they are not mounted. The output is fine with equal spaces between each column (like between col 2 and 3.. or 3 and 4). I want to do a (dare I say) sed or awk or something to reduce the space between 1st and 2nd col. So that it has space like between col 3 and 4.. or 2 and 3. This is because I am showing this output somewhere and because of the space its not showing up correctly.Also I hope the command will still work when the NAS drives (afp) are not mounted.. basically consistency. The spaces are not showing properly in the quote tag. Changed it to CODE tag.
I tried playing a DVD several times and each time, after opening in Movie Player on Ubuntu 9.10, it froze up, I forced quit and then I noticed a huge reduction in free disk space on my root (/) directory. Has this happened to anyone else?I figure I Movie Player probably saved some temp files somewhere on hard disk and wasn't able to clean them up since I had to force quit. Any thoughts about how to clean this up and take back that previously free space?
A few days ago, I got a message that stated I had zero bytes of disk space left.Odd, I thought, but I had been doing video transcribing and thought that may be the issue.I moved a video (4 GB) off the hard drive to an external drive and then went about my business.This morning, I got the message again. I enclosed a screen shot. I moved a few more items off my hard drive - but then was soon out of space again. (Less than an hour later.)I logged in as root and poked around. I noticed that /var/archives had almost 60 GB of data in .tar.gz files.I moved them off to an external drive and am okay for now.
I have a dual boot system that only has about 6.5 GB of total file space for Ubuntu on the disk. Recently I upgraded to 11.04, and have had problems logging on and in downloading and installing programs. Occasionally I get messages that say available memory [edit: I meant disk space, not memory] is too low.
I see from the disk analyzer that a folder called tmp is very large. Can that file be safely deleted? Anything else to clean up and scavenge more space?
I'm a little puzzled as to why I'm getting a warning about running out of disk space. It seems others have similar issues but with little resolvI received a warning about how I have little disk space remaining. I got the message when writing files to my /home directory.The output of df -h is:
Code: Filesystem Size Used Avail Use% Mounted on /dev/sda1 7.4G 5.9G 1.1G 85% /
running ubuntu 8.10 amd64 on my HP intel 2.26ghz dual core processor, 320gb hdd and 4gb ram. My main problem is that when I dual booted ubuntu onto my laptop, I had vista already installed, so I only set aside 20GB of hard disk space for ubuntu, which seems like more than enough to me. What's going on is that it says I have used all my space and only have one GB left, when the only thing I have on my computer are word files and necessary installs like flash player, compiz, avant, etc.When trying to hunt down what was using all my space, first I checked out the disk usage analyzer (some pics from here in):theres my hdd with 1/19 gb remaining here's a breakdown of my home folder, showing that it's not using very much at allhere's my filesystem, now we can see that all the space is being used in the folder: /var
I seem to be running low on disk space on my linux server. 'df' shows about 82% usage on a stock CentOS install with sendmail. I routinely delete old email directories, but for some reason, I stay pretty high in disk usage. Is there a fancy little bash script or something I could run that would find the biggest files and I could go get rid of them?
I use Linux. I coded a screenshot program some time ago and now I have 9 GIG of screenshots, 60000 JPEGs, most of them look pretty similar, and I have 300 MB of disk space remaining.
What are some good ways to start to compress batches of them (or all of them) in the background given the limited space? The problem with compressing the folder all at once is that I wouldn't have enough disk space for that. It seems the process needs to be broken down into chunks. So maybe something like: Get a list of all the files Add a chunk of the files (say, 20) to a compressed archive. Once it is done and saved successfully, delete the chunk of files
My main storage partition got full, so I'm deleting files to make room. However, df -k keeps reporting no space available on that partition (/disk). Here's the output of the command several minutes apart while another process is deleting a 30G of space:What can I do to make the space available immediately?
I used photorec to recover lost files and it brought up 70gb worth of files, when I was done looking through them I deleted these files. but these files still seem to be taking up my disk space. When I try to access my trash bin with root I get a message that reads...."The folder contents could not be displayed. sorry, could not display all the contents of "trash": operation not supported." if I open my trash bin when I'm not in root, the bin is empty.
I am trying to install CentOS 5.2, and the installation ran out of disk space after running for about 2 hours.I checked the FAQ, and it said 1.2 GB. The disk is 3 GB. The default install was selected, and I think that it checks for sufficient available disk space before installing. Still, it ran for quite a while before announcing that it was out of disk space.The Installation Guide is not very helpful, since there is a blank page where the disk space requirement is supposed to be. I just picked the default installation. A search of the forums on "not enough disk space" did not return much.
I am a both rhel5 and fedora user.I can not configure my Samsung syncmaster 632nw monitor to display full screen at 1360X768.There is huge black space both left and right side of the monitor. I have tried many times to solve it but unable to do it.The max screen resolution is 1024*768 and minimum is 640*480.
Running ver 9.10 with Wine installed. Up untill about 2 mos ago Wine loaded 'Rosetta' ok. I tried last week and get Application Error. "Cannot create needed files of not enough disk space". I have over 35 gb unused space. I removed wine (purged) and re-installed -- still no good. It however works on my laptop that has same configuration (9.10)??
I installed Ubuntu 10.10 with wubi and i have been enjoying my Ubuntu experience a lot. I installed quite a bit of programs and spent a couple hours customizing my machine. The problem is im running out of disc space. Any ideas on how i can add more space. I have gparted but i dont know where to move the free space to because wubi installed it.
Today I was installing a lot of software since I'm just setting up my Slackware system again after a fresh install, and I realized that my root partition has very little space left.
Here is the output of df -h:
Code:
As you can see, I have a 20G (19G here for some reason) root partition, 8G /var, and 86G of /home. I thought this would be plenty since many recent recommendations for / are 10-15G. Now, though, 17G are used up for some reason! How is this possible? I thought a full slackware install only had about 4G of software! I don't have any music or movies or any crazy huge files that I know of, and those would be in my /home directory anyway. Is there any way I can see which files are taking up all this space?
If it's necessary to allocate more space to my / partition, is it still possible to boot up a GParted live Cd, shrink /home a bit, move some partitions to the right, and expand my root partition? I would REALLY prefer I don't have to reinstall since I just spent a ton of time setting up my system again, but if worst comes to worst ... :'-(
While installing OS, in partition window after OS file system structure I've left 277 GB. But after installation it shows Size - 255GB and available disk space is 242 GB.
Isn't it weired? How can I use the total amount of space in Linux? I need the whole 277GB exactly. What should be my workaround?
Some thing is using up a huge amount of my disk space about 10G and I can not determine what it is. When I look at my disk usage in system monitor it say I have used about 25G and when I scan the directory in disk usage analyzer the entire file system used is 15G.
ran out of space in my /home dir. Have a second hard drive to install and would like to designate it as additional space for /home. I do not want to mount it as a dir inside my home I would like it to simply work as though my /home simply has more space available to it.
I'm using Fedora to program Android, and have a directory with all the files I need. But my home directory isn't big enough to store my android directory. Is there any way to make more space to do this?When I have tried to install something using "sudo apt-get install (program name here), it says my name isn't in the sudoers file. What is this and how do I add my name to it to be able to install things from there?
I have a 60GB partition with / and home on it. I logged on yesterday and it gave me a warning saying that I had only 1.9 GB of disk space left. I ignored it for a day and assumed that i had too many videos and pics.But the next day i had not added any files or downloaded any software but i had 0B left. I used the disk usage analyser and found that 33GBs came from /var/log. It was from two log files. syslog and daemon.log 16.5GB each!! I opened them up and i found that this line of text was repeated hnundreds of thousands of times.
Code: Jul 22 19:32:36 aulenback-desktop ntfs-3g[5315]: Failed to decompress file: Value too large for defined data type
i have about 2 TB of 700mb avi files as data on disc want to spread it across two 2TB ext usb drives (sata 3.5 inside the housing) obviously i have to rip them to the laptop and then move to the ext hdd (omg laborious little task) am i better doing the ripping in meerkat or in a windows machine? files need to be accessible by W7, XP, and meerkat to vlc player. what should i format the discs to?
I'm looking for a fast way to verify a copy of a folder with 150Gigs of data, in 33 files. Some of the files are a few kb, while a few are 20-30Gigs. I've done a file count, which is quick, but doesn't verify that all the files are intact. I tried running md5sum on them, which works, but will probably take as long as copying the files in the first place. Diff works too, but is slow too.