But, when I run the du command, it shows I have used only 2G out of 9.9G.
ip-10-204-70-44:/$ du -xh --max-depth=1
14M ./etc
4.0K ./mnt[code]....
It just driving me crazy and interesting too. This is big problem for us since the root disk / is full and some of the function in our site is failing.
I have been using Debian on and off since Sarge and it's not as long as most people but since I can remember, the installer has always had some issues and I was wondering if you guys who use Debian religiously could answer these questions for me:
1. Difference between "Netinst" & "Business-card" besides the size being smaller? It seems like the Netinst installs two different kernels. One dated kernel to get the system up and running and then it downloads the latest one where I think the 'Business-card' ISO simply downloads the latest kernel and nothing more which seems preferred. Am I wrong?
2. After I install using "Netinst", why when I immediately perform an 'apt-get dist-upgrade' are there multiple packages to upgrade? I thought the whole point was to install fresh updated packages, no?
3. Why has nobody fixed the installer progress bar from hanging on 33%? As far as I have been using / installing Debian, the installer progress bar is always stuck on 33% & it still installs fine but never shows the user it's exact progress. I would think the developers would have resolved this after 4 years by now...
my question is according to google search I need to umount the filesystem point in order to reduce its size.... so it means I will require that nobody uses the system? since /var logs a lot information?
I also have the option to reduce /root but I think is more complicated.
so what are the precautions in order to reduce /var/?
When Installing Ubuntu what difference does the installation size make? Right now it's set at 17GB :/. The higher the install size better functionality or what?
We use VxVM and VxFS on HP-UX and Ie used them in the past on Solaris.ve found they are available as Storage Foundation v5.1 for 64 bit RHEL.(In fact there�s even a BASIC version for free on 2 processor systems). Previously wed run into a 2 TB limit for filesystems on the older versions we have on HP-UX. The data sheets at Symantec are pure marketing fluff. Does anyone know what the filesystem size limit is for 5.1 on Linux?
How can I shrink a partition to the filesystem size?I copied my ubuntu installation via dd from a smaller partition to a little bit bigger one to have all the same settings and programs and upgraded the distribution afterwards. Now the filesystem is smaller than the partition size. It would be nice to have both partitions have the same size so I could copy back the newer distribution someday ...GParted recognizes the new partition as 27 GB, the filesystem is just 25 GB.Is there a nice way to resize the partition to exact the size of the filesystem so that the filesystems remains untouched and no data might me lost?
Question:Several months ago I upgraded my hard drive on a dual boot machine (xp & U9.10 - both 64bit) from approx 80gb to 250gb. I chose to use DriveClone from windows to clone the old drive directly to the new drive. Operational everything works fine. the problem is the Ubuntu HD size went from around 10GB to approx 150Gb, but the filesystem only recognizes approx. 10GB as the total available capacity. I just tried to reinstall Ubuntu without reformatting and only recovered a few GB's. GParted reports the proper partition size. Is there any way I can correct this problem? It's not a big issue...I thought I would check here first before I reformatted the partition. The pre-boot disk check ran and reported no errors but the file system is still not seeing the unused GB's.
We have 3 RH5u4-64 servers. Server 1 is a standalone server. Servers 2 & 3 are clustered filesystem servers running Veritas CFS 5.0mp3.
Server 1's filesystem is EXT3 and was cloned from a Sun server running Veritas 5.0mp3-VXFS. Filesystem size returned from 'du' 'df' show about 428GB on both the Linux Standalone Server(EXT3) & the Sun Solaris Servers (vxfs).
We then cloned Server 1's filesystem (EXT3) to the 2-node CFS servers. Cloning was successful, but the filesystem sizes returned by 'du' 'df' show 128GB. Block Size for the EXT3 filesystem is 4k while blocksize for the VXFS filesystem is 1k.
Where did that other 300GB go?
I can see VXFS/CFS being slightly more efficient than EXT3 because it's been around much longer, but that can't possibly account for the vast difference.
What originally seemed like an easy thing to calculate has given me a big headache. Perhaps someone can help me with my issue. I am trying to find, in particular, how much memory certain application processes are taking. The process always is the same name, main_server, but with an argument to tell it what to do when running as a daemon.
When running the following command against all "main_server" processes, it produces a result in megabytes based on the output of the rss field in 'ps'.
Code: CALC=0 for ea in `ps -e -orss=,args= | sort -b -k1,1n | pr -TW$COLUMNS | grep main_server | grep -v "grep main_server" |
[Code]...
Currently I am left scratching my head. For capacity planning purposes, it would be nice to know how many more 'main_server' processes could run on the system without causing it to swap. Knowing the buffer and cache usage will go down as running processes demand more memory, I prefer to look at the free memory excluding cache and buffers. However, since 'ps' is reporting the processes are using more memory than free reports is in use without those things, I have no way to know how many more processes the system can support. I played around with different fields in 'ps', such as vsize, size, etc, but with no luck in matching up any numbers.
I'm trying to create a script that will find the bandwidth usage of certain protocols only. For example, SMTP. I would like it to just return a number. Is there a known command/parameters to output something like this?
I bought new hdd (WD2002FAEX-007BA0). This disk have normal block size (512bytes) so I do not bother with the alignment.I trying make test write speed.
Code: dd if=/dev/zero of=/dev/sda oflag=direct bs=16384 count=100000 100000+0 records in
I am using Fedora core 10. I have changed my partition size of Linux from windows. After I finished resizing the partition, I rebooted my system to the Linux platform. While booting it gave me an error: repair filesytem #1: I don't know what to do?
I am running CentOS 5.5 with a 14T ext4 volume. We are sharing out a few sub-directories via NFS. Our customer was doing some penetration testing with our web app that writes to one of those NFS shares. We are not sure if they did something to cause the metadata to grow so large or if it is corrupt. Here is the listing:drwxrwxr-x 1 owner owner 470M Jun 24 18:15 temp.badI guess the metadata could actually be that large, however we have been unable to perform any operations on that directory to determine if the directory is just loaded with files or corrupted. We have not run an fsck on the volume because we would need to schedule downtime for out customers to do so. Has anyone come across this before
i have installed ns 2.33 and added the patch mobiwan for supprting mobile ipv6. i ran some tcl scripts , and below is the trace file generated. calculate the handoff latency, and if someone has awk or perl script to do so, tcl file [URL] out.tr [URL]
I recently installed Bio-Linux 5.0 as a dual boot system with XP for some bioinformatics applications, but Im having some problems with the amount of disk space which can be allocated specifically for the Ubuntu install.
Ive been using blastclust to analyse some very large data sets, which keeps on crashing due to filesystem running out of disk space.
When I installed Bio-Linux 5.0 from the live cd, the maximum size I could allocate to the install was 30 GiB, and I havent been able to find a way to change this.
Ive tried using System->Administration->Partition Editor using the live cd, and can view / delete the partitions, but I cant find a way to specifically alter the disk space allocation for Ubuntu.
How do I increase the filesystem size to larger than the current 30 GiB?
Actually i m new to scripting in Linux, we have RedHat Server Edition 4 in which different software licences are installed and software itself installed at client side. the usage of the software is recorded daily in a text file which is of format
I want to write a script that checks that for any OUT & IN, if utility, user and Client PC Name are same then subtract OUT Time from IN Time. This has to be done for each & every user and utility.Is there any GUI programming possible in Linux.
I am new to shell script and to this form as well, I did try to search for a similar post like mine here, but could not find one.
Here is what I'm trying to do:
I am trying to grep server logs to find a specific string and then capture the time stamp and the value of that grep string in them. The log file prints out messages on per sec basis.
My script is able to grep the server logs for the entire period of my load runs and then outputted it to a .csv file too.
Unfortunately this .csv file is too large to extract it on my PC and to generate graphs as it exceeds the excel limit. I need some help on how to read this .csv file in a shell script and then take an average on per min basis before I can export it out on my desktop and generate graphs for analysis. example of the out in my .csv file:
today I upgraded via official testing repository Gnome to version 3.18. After this, icons on desktop and nautilus are bigger, than before. Next thing, gaps between icons are smaller than before. I tried change theme to default (Adwaita), then run gtk-update-icon-cache, but without result.
Normal view - icons are big for this view. URL....
Small view - icons are still big for this view. URL...
How can I change icons size and gaps size? Or is it bug for this version?
I've had a look at some similar threads but as I'm very new to linux they're already a bit technical for me. Sorry, this calls for someone with patience. I gather from other threads that disconnecting an external drive without unmounting is a no-no, and this seems to be the likely cause. Now the disk is read only and I'm unable to change any settings through the usual control panel on ubuntu. I'm just not familiar with the terminal instructions. I tried to cut and past a few command lines from other threads but I got some warnings that proceding could damage data. Like this one: WARNING! Running e2fsck on a mounted filesystem may cause SEVERE filesystem damage.
I am very new to linux, and I have a question regarding the filesystem check (fsck). The power recently went out and when I tried to restart linux the following error appears:
*/dev/sda1 contains file system w/errors, check forced it then goes on to say..
*An error occured during the file system check. Dropping you to a shell; the system will reboot when you leave the shell. Give root password for maintenance (or type Control-D to continue) I wasn't sure what to do, but checked some other online forums and they suggested running fsck manually - so I typed in the root password - and used the command, "fsck -A -V ; echo == $? ==" it then gave the following message
*WARNING!!! Running e2fsck on a mounted filesystem may cause SEVERE filesystem damage *Would you like to continue (y/n)
Again, I wasn't sure what to do so i just checked no. I then manually turned off the computer and was prompted at the beginning to press Alt-3. I was brought to another screen and it informed me one of the drives was degraded and suggested rebuilding the array. I tried doing this, but it still brings me back to the original error of, "/dev/sda1 contains file system w/errors, check forced," and the process continues.
Also, when I tried to rebuild the array, I didn't backup any of the data on our home directory before doing this (which was probably a big mistake). After being prompted to type the root password, I was able to give the ls command and look at all the directories...the home directory where our data was stored was empty and I am afraid I may have lost some information. Is there a possibility that data was lost when I was trying to rebuild using the old drives?
I have an array called arrayini which stores numbers. I want to take log to the base 2 of each of the numbers in that array and put it in file called result. I've used the following code to do it.
Code:
size=${#arrayini[@]} for ((i=0;i<size;i++)) do echo "scale = 12; l(${arrayini[$i]})/l(2)" | bc -l done >result
It works fine but its taking pretty long to calculate since I've got about 230,000 items in the array. So I decided to store the result into an array hoping that it'd be faster. I tried the following code. arrayresult is where I try and store the result. The code doesn't work because of the second last line.
Code:
unset arrayresult size=${#arrayini[@]} for ((i=0;i<size;i++)) do arrayresult[$i]="scale = 12; l(${arrayini[$i]})/l(2)" | bc -l done >FILE2
I notice that on DistroWatch, the descriptions include "debian based", "arch based", etc. For the newbie, is there any practical difference between the different bases, or is it a Coke vs Pepsi type question?
Is a distro based on Debian say, easier for a newbie to learn or work with than say one that is FreeBSD based?
I have read many articles on hdparam to calculate the disk read and write speeds and some on interface and CPU limits. But is there a structured way of calculating the maximum throughput of a server including all the subsystems. Like storage, CPU, network, memory and so on? So that I can create a script that i can run on a newly installed Linux machine and calculate the maximum throughput .
calculating the IOPS for my system ?I had understood that IOPS could be calculated by monitoring IOSTAT. The r/s (reads/sec) and w/s (write/sec) combined basically gave you the peak IOPS.However having monitored iostat for my app during a performance test, the figures I'm seeing lead me to believe that iostat couldn't be a reliable way to calculate IOPS. I get values of over 10,000 for w/s! I didn't think this was possible given the storage system i have (4 x 7500rpm sata disks in raid 5).