General :: Fscheck For Ext3 Filesystem - Usually Occurs When Reboot System Performing Administration Tasks
May 29, 2010
fscheck is quite annoying, since it usually occurs when I reboot my system performing administration tasks. do I actually need fscheck if Im using the ext3 file system? if not how would I extend the period between checks or just turn it off altogether?
An error occurred during the filesystem check. Dropping you to the shell; the system will reboot when you leave the shell. Warning -- SELinux is active Disabling security enforcement for system recovery Run 'setenforce 1' to reenable
I am running Ubuntu Linux 10.04.1 on an AMD 64-bit system. The server is primarily used for Windows file sharing via Samba in a small local network. I use webmin and putty to administer the system. I have two 1.4 TB drives for storage and one 500 GB drive with 18 GB mounted for root.I performed a large cut & paste operation (25.8 GB of files) using the File Manager in Webmin to move a certain folder into another folder within /media/Work. The operation failed and I am now getting a "root filesystem full" error, and am stumped.
I've ran fsck -c on the (unmounted) partition in question a while ago. The process was unattended and results were not stored anywhere (except badblock inode). Now I'd like to get badblock information to know if there are any problems with the harddrive. Unfortunately, partition is used in the production system and can't be unmounted.
I see two ways to get what I want: Run badblocks in read-only mode. This will probably take a lot of time and cause unnecessary bruden on the system. Somehow extract information about badblocks from the filesystem iteself. How can I view known badblocks registered in mounted filesystem?
I have installed ubuntu to my pc. i made 3 partitions. one for system, one for data and one for swap. two of them were ext4. after some time i have reinstalled ubuntu again. but this time i didn't put to format the second partition, but just mount it using ext4. after that i cannot open my files. checked with gparted shows that 2GB used, but with df 188MB. and in properties writes ext3/ext4 filesystem. i used chown, chgrp but didn't help. please help, these data are ver important. i cannot lose them.
I often get responses from people who first say: "Are you sure? You want your network to be exposed to the outside world?" I am not experimenting on a Production Server of NASA or any Security Concern Department. Friends, there is no harm in experimenting on your personal computer or on a test computer which is isolated from the production environment. Look at hackers! What do they do? If they don't know how security is breached then how would they come up with security measures?
If my question reads... "How to let any user perform Administrative Tasks on a Linux System irrespective of his/her privileges on that particular system?" then I would not get the right answers in the first place. They will say... "You are letting everyone destroy your system... are you sure you want to do that?" My question is: Why should we restrict ourselves from experimenting even if it sounds weird to other people?
I give you an example where it is desirable to let an unprivileged user perform certain tasks. You want to know if there are any employees in your office who are storing videos in their home directory and filling up the disk space to a great amount. You have a department called "Command Center or Data Center Operations or Help Desk" call it whatever you would, whose work is to monitor such activities, and you create an account "monitor" for them to monitor such activities but they are not able to do them:
I am trying to figure out a totally odd behavior of the ext3 filesystem mounted in Ubuntu 9.10. There is a Korn Shell script, part of which does the following in the loop:
while ((1)); do mv dir1/file dir2; if [[ ! -r dir2/file ]]; then echo "ERROR" ls -l dir1/* dir2/* exit 1 elif echo "OK" fi done
Given that dir2/file always exists and that I do not move it asynchronously with "&", my script should never hit the "ERROR" statement. The odd thing is that it does, and quite randomly (no pattern at all). However when it does hit the ERROR case, ls -l prints that file is in dir2 and it is readable! I tried using "-e" instead of "-r" test - no luck. I never seen anything like this in 10 years of my programming experience. Same script worked fine on Fedora 11, and yet it wouldn't work on Ubuntu.
I have an external 3.5" USB 250Gb HDD which is showing symptoms of hardware problems (repeated /var/log/messages errors of "reset high speed USB device using ehci_hcd"). This was originally plugged in to my NSLU2 running Debian Etch. I have just installed Ubuntu Desktop 9.10 to a spare Pentium-3M laptop and was hoping to copy the contents of this HDD to a fresh drive. However, I cannot mount it even read-only; mount -o ro /dev/sde3 /mnt/disk fails, and the /var/log/messages error is "recovery required on readonly filesystem", "write access unavailable, cannot proceed". I cannot understand why mounting a disk read-only should require write access. Following advice I googled elsewhere, I tried running mke2fs -n /dev/sde3 to try to list the alternative superblocks - but once again I got the error that the device was read-only. How can I go about accessing the data on this disk?
It is just a general discussion on "How Can Perl Be Useful in System Administration and Application Development on UNIX and Linux Platforms"? I am doing pretty much well with Bash and PHP. In fact, I am learning and experimenting in depth in those two domains. The project that I am working for these days has several scripts which are written in Perl and Ksh. They interact with Ab Initio graphs and perform extensive jobs. Anyways, my current role doesn't require me to have knowledge of Perl but I don't see any harm in learning something new.
So, before I take it, I want to know how Perl can help me and what I can do with it as a UNIX / Linux System Administrator and Application Developer? I just looked at some of its keywords, functions, and commands. Perl Variables are identical to those of PHP. Conditional Statements, Loops, etc. are also much the same as found in any other computer languages. So, at first glance, I have got this feel that it wouldn't be much difficult for me to learn it. But what do you say? Do I really need Perl when I am focusing on PHP and Bash?
I'd like to know some good resources, links, e-books to learn basic Linux system administration scripting. I know how to write very basic BASH shell scripts so I'm still at beginner level. There are so many docs out there that I'd like to narrow down people's opinion's on the best easy to understand resources that are available. Eventually I'd like to learn basic PERL as well.
When installing fedora 10 from scratch on an acer aspire one 150L, which filesystem should be used? ext2 or ext3? a basic explanation of the reason would be great too.
I was running ubuntu as a live cd and I wanted to format my pen drive into an ext3 filesystem. I put in sudo mkfs /dev/sda1, but know im thinking that sda1 was my HDD!! I removed the cd from my computer, and it wont boot up into windows anymore!The only thing that is giving me hope is that the mkfs took about 1 min to format whatever it was formatting (my pen drive or my hdd!!) and my hdd is 500gb big. Is there anyway that I could have accidentaly formatted my HDD?
I run an upgrade and an update on a lucid lynx beta 2. --- got no problems. but about the filesystems i have some questions because it seems for me that at every system boot the system will run an fsck. somtimes it's shown up, somtimes not. but in /var/log/messages and in syslog
I have always following messages ( occured in beta 2 too ).
But first before i continue - here my disk layout:
And here my filesystem types:
This is my problem because those values are seems to be static ! ( note: this partiton is mounted but not in use ) and last not least: the drive is an external usb scsi disk. but on the other side lucid lynx is running fine on my box.
I was looking in a dir I haven't used for a while (I use it for data storage) and found a directory was repeated - that is, two directories with the same name. I renamed one of them, on the assumption that one of them has a non-printing character in its name, and that worked without a problem, but ls -i lists their inodes as the same. ls -l says they both have two inodes (. and ..), and the files they contain have the same names. The inodes of the files is the same (that is, the inode of a file in dir A is the same as the file of the same name in dir B.) Each file in each dir is listed as having only one inode, but it's the same as that in the other dir.
I'm wondering if anyone knows what will be Squeeze's default filesystem. Will it be the proven ext3 or the newer (sometimes faster, sometimes slower) ext4?
I currently have ext4 and I have nothing to complain about. In fact, my overall experience has been very positive. Ext4 is definitely faster when fsck runs during boot.
What would be the cons of having ext4 as default in Squeeze?
I have a external HDD with eSATA and USB connectors available. I want to use this HDD to store my backups. The HDD should be encrypted (my main system is as well).
So here is what I did so far: 1) I used the following code to create the encrypted LUKS partition with EXT3 Filesystem: Code: cryptsetup -c aes-xts-plain -s 512 luksFormat /dev/sdb1 cryptsetup luksOpen /dev/sdb1 luks mkfs.ext3 /dev/mapper/luks The system always hang when I executed the "mkfs.ext3..." command, so I switched the HDD from eSATA to USB and then it worked fine.
2) When I switched on the ext. HDD the first time, the drive was recognized automatically and Nautilus asked for the password. I typed it in as checked the checkbox to remember the password in the future. For the backup I use a nice script that I found in another forum, where I can define a mountpoint and then the script will check for previous backups and only make a incremental backup based of the latest version. The script also mounts the drive automatically. In order to always have the same mountpoint, I want to make an entry in the /etc/fstab using the UUID of the ext. HDD.
Whatever I tried, it doesn't work. What am I doing wrong? Here is my current /etc/fstab Code: # <file system> <mount point> <type> <options> <dump> <pass> proc /proc proc defaults 0 0 # / was on /dev/mapper/ubuntu-root during installation UUID=2ea47421-73ce-4c66-9606-8a1db81ae640 / ext3 relatime,errors=remount-ro 0 1 # /boot was on /dev/sda1 during installation UUID=dbdeb793-1d4e-43ea-8986-7b37fdbc9674 /boot ext3 relatime 0 2 # /home was on /dev/mapper/ubuntu-home during installation UUID=42702091-83e6-43eb-aad1-108f43eedf9d /home ext3 relatime 0 2 # swap was on /dev/mapper/ubuntu-swap during installation UUID=e225bcf9-908b-4226-a963-6b02ee658df1 none swap sw 0 0 /dev/scd0 /media/cdrom0 udf,iso9660 user,noauto,exec,utf8 0 0 # Eintrag wegen iPhone none /proc/bus/usb usbfs devgid=125,devmode=666,nodev,nosuid,noexec 0 0 # external HDD UUID=913977f7-8fa6-416f-af79-b5f913b68f53 /media/backup-hdd ext3 noauto,users 0 0 I made the "none /proc/bus/usb..." entry because it was recommended to ensure correct behaviour of the iPhone. Not sure if I need it though.
I created the mountpoint with this command: Code: sudo mkdir /media/backup-hdd Now it seems the mountpoints owner is not root - strange right? Code: 2 4 drwxr-xr-x 3 michael michael 4096 2010-01-15 02:45 backup-hdd How should I mount this drive correctly? It will be automounted as every USB device, but that should not be the case. I want the script to mount and unmount the drive.
I just got a new hard disk so that my "/" and "/home" partitions would be located at their own separate drives. All was well until i tried to expand my "/home" partition to fill up the entire drive that used to also have the "/" and swap partitions on there. Let me sketch the before-and-after scenario in GParted,
Before:
What actually happened after:
Gparted gave an error prompt after which i found out that the entire ext3 partition (/home) had been moved to the left but not yet expanded to the right, which is of course not a problem. However, for some reason it has labelled the entire ext3 partition as a swap partition!
I'm still running ubuntu from the usb flash drive, because i dont want to risk that the (120GB) swap area will actually be used and cause my data to be lost.
I guess my question is, can i relabel the swap partition as ext3 like it was before without formatting (ie without losing data)?
Here's the output of "fdisk -l", which doesn't show the swap area (the drive in question is "/dev/sdb"):
We have 3 RH5u4-64 servers. Server 1 is a standalone server. Servers 2 & 3 are clustered filesystem servers running Veritas CFS 5.0mp3.
Server 1's filesystem is EXT3 and was cloned from a Sun server running Veritas 5.0mp3-VXFS. Filesystem size returned from 'du' 'df' show about 428GB on both the Linux Standalone Server(EXT3) & the Sun Solaris Servers (vxfs).
We then cloned Server 1's filesystem (EXT3) to the 2-node CFS servers. Cloning was successful, but the filesystem sizes returned by 'du' 'df' show 128GB. Block Size for the EXT3 filesystem is 4k while blocksize for the VXFS filesystem is 1k.
Where did that other 300GB go?
I can see VXFS/CFS being slightly more efficient than EXT3 because it's been around much longer, but that can't possibly account for the vast difference.
I am going to install CentOS 5 on my pc , for3d modeling , and i have some doubts about partitioning the system.
I know that i can have /, /var , /tmp , /home and other directories as individual partitions. I know that i can use ext3 , or xfs ( i know that anaconda doesn't let to use xfs, but i know a way of migrate partitions to xfs after installing) or ext4 , and others , without problems.
But i could for example, once installed all with ext3 , have for example / on ext3 , and /var and /home on xfs.
For me it would be very interesting have /home in a another partition and as xfs (migrating it later from ext3) as mostly i will save in /home a lot of big files ,images,dvd isos,renders , binary maya files and like, and i see a advantage of use xfs with /home.
But a lot of people also use xfs for / as well , but as / has a lot of small files , perhaps use xfs on /var and some other directories, could be good (because xfs performance deleting files or other features), and leave / as ext3 for example.
So is it possible to have different filesystem type for some directories under / ,(for example /var and /home as xfs, the rest as ext3 in / ) , without having any problem ?
(not counting using xfs for a directory that won't see improved his performance because has a lot of static small files or like)
I recently moved around, deleted and just had a general day of cleaning up the file structure on my pc. HDD contains about 2 TB of data. Around 1 TB was relocated on the disc, and yet another TB was moved to external drive.
In windows this would likely means it was time for a defrag. Can't find any defrag tools in Xubuntu and that lead me to remember I read somewhere that there is no need to defrag when running Linux.
So is that correct? Is defrag only for the wondows world and something we never need to do in Linux?
If not is there a way to do a defrag from terminal?
Is that possible, I mean when I upgrade F10 to F11 with yum upgrade is there a way to 'upgrade' the filesystem to ext4 for example (with the exception of boot partition)? Or I have to reinstall fedora like new?
While changing the filesystem can I do it by parts? what I mean is for example: I have 2 partitions like '/' and '/home' with ext3, so I backup data in '/home', change '/' to ext4 then mv files from '/home' to '/' and change '/home' to ext4 and finally mv those files from '/' to '/home'. Is that possible?
I have two ext3 partitions within an encrypted lvm2 volume. when i start up my system it says that there are 0.3% non contiguous blocks.
This is my steup:
When i want to repair with repair system from dvd it tells me that the repair and check operation for encrypted LVM devices is not supported. so how can i fix my filesystem?
At boot time, before entering Runlevel 3 the HDD will go mad when mounting tmpfs on /dev/shm...
Code:
EXT3 FS on sda1, internal journal EXT3-fs: mounted filesystem with ordered data mode /dev/sda1 on /boot type ext3 (rw) tmpfs on /dev/shm type tmpfs (rw) INIT:Entering runlevel 3
It will go on and on at the tmpfs on /dev/shm type tmpfs (rw) until i press ctrl-C...then I will stop whatever it is doing, let the hdd rest a bit, and resume normal boot..
When I rebooted my system (RHEL 5.3 x86_64), I got the following error when it was booting up:
Code:
Full screenshot of the bootup error messages here: previous bootup error
So to try and fix this, I booted from the OS disk and did a linux rescue.
Here's some of the results of the commands I did:
Code:
Here's a screenshot of the output from fsck after it's done, notice it says 0 bad blocks but the system was modified:fsck results
I then reboot but now I get the errors:
Code:
Full screenshot of the startup errors I get now here: current bootup error
Does anyone have any ideas on how I can fix this? Was the fsck I did initially the correct thing to do to fix that initial group descriptor error or did I make things worse?