I have an NTFS file system nfs-automounted on our RedHat servers. Users can read and write to the file system no problem, and can create new files, edit them, and delete them to their heart's content. The only issue is that utilities such as "dos2unix" cannot create temporary working files:
$ dos2unix events.0818.dat
dos2unix: converting file events.0818.dat to UNIX format ...
Failed to open output temp file: Operation not permitted
dos2unix: problems converting file events.0818.dat
This isn't limited to "dos2unix"; any other utility that creates a temporary working file gets the same problem. If I copy the file to a local file system like /tmp, it works fine. Here's the kicker: this works fine on Solaris systems. I can take the "dos2unix" utility over to a Solaris system that has that exact same NTFS file system automounted via NFS, and it works. No issues creating temporary working files at all.
I have a shell script that need to create some files:
1) backup files of user passed in file ( that will be written by this shell ).
2) temp files that the shell will create and later delete/remove.
This shell script will be used from my local dir ( I am not a super or a sysadmin ). Users of this shell will call this script to run on their local files in their respective directories. When my script runs, it errors with the following:
There were some files residing on my ext3 file system, using Ubuntu as my linux distribution. Yesterday I formatted the hard drive using a windows install CD, rewriting it with a new NTFS partition. I'm willing to restore my personal files deleted due to this format.
I am using centOS-5, I have mount NTFS drive by using fuse. But there is no rights and even there is no option on right click to make new directory or to del any file or folder. This is line of fstab for NTFS drive
How can I get full access and control on this NTFS mounted drive.
I want to back up an entire Linux system on a 3Tb external Western DIgital USB3 drive.
I do not want to reformat it from what it is, apparemtly NTFS.
Is there a utility that can act like a file manager like mc, that will permit me to create an ever expanding (to 320Gb) TAR file that will retain all the original file permissions. I have had nothing but disappointment with Linux backup utils with a FAT32 external drive, and I am concerned if I just try an tar the entire drive at once, with around 3 million files, I might run out of memory.
I know that Linux is open source but there must be ways of creating non-open source programs to be run on a Linux system? Does such a thing exist and/or have a name? Would any source code that has been compiled be unable to be read by anyone properly unless the soruce code was released?
I have an ntfs partition that I wish to access as a normal user(non-root). For this I did the following. As root I created a folder /windows and did a chmod 777 -R on /windows. Then I added the following line to /etc/fstab
Now, the partition is mounted alright but the problem is that when any other user (non-root) creates a files in /windows (say by executing touch newfile) the newly created file has the owner and group set as root. The non-root user can create the file and he can also delete the file, however, he cannot change the permissions of the file and also the owner:group is always set as root:root. How do I get across this problem, i.e. how do I mount a partition, so that a non-root user can also change the permissions and ownerships of the files he creates.
I am using back in time to back up files from home and from another mounted directory on my system (ntfs). The back-ups are occurring automatically and appear to be complete; but, I cannot delete old back-up snapshots in the backintime GUI Also with sudo nautilus or as root in terminal with (rmdir) I cannot delete the snapshots. My drive is filling up and rather than uninstalling back in time, I would like to simply delete the unneeded snapshots. How can I delete these files? Is there an rsync file that I should configure to delete these? My expectation of backintime was that it would back-up at the requested frequency and not create complete duplicate copies of the files, but, use symbolic links to unchanged files. How can I verify if this is the case? Does the cron file control this>
A non techie friend has helped an even less techie friend by contacting me by email to discuss an ailing laptop. A few emails were exchanged, with more details, and it was not looking good because it seemed that suddenly the CD drive was not responding, nor any USB devices, the wireless icon was gone, but Ubuntu still seemed to work (for now), with wired ethernet also working. I was struggling to think of what could be done, with the favourite routes of Live CD and Live USB apparently gone.
After a few more hours - another email: 'It's now working! After so many reboots it checked disc for errors and repaired itself! Is there some way of doing that when needed anyway?'I see there is 'Disk Utility', and this would presumably fit the bill, but how does it do checks and repair when the damaged file system is being run, and is currently *mounted*? I thought utilities like fsck(?) could only be run on unmounted file systems? Have I misunderstood the disk utility fs check repair function? And anyway, what might be a good answer to my (nontechie) friend's question 'After so many reboots it checked disc for errors and repaired itself! Is there some way of doing that when needed anyway?'
For the record: (quote) It is a toshiba EA60-155 Model number PSA67E-00300C8J. He put in extra ram to install ubuntu. He thinks he may have deleted something! There is a 'trash' file on his USB drive with loads of stuff in it and he doesn't know how or why but because it won't now read the drive on her laptop we cant replace it! (end quote)
a friend of mines laptop died a few weeks ago and i said i would extract data off the drive for her... did this with dd and created a .image file. I have this image mounted so it is browseable, and we can copy files from it however i am unable to delete files from it?
You'd think it was not a big deal... but i'm running out of space!! and the .image is 90 odd gb. I know i "could" just get the disk and copy the files off she wants... but that would be too easy
I am using Gentoo Linux and for a while now, the root file system is mounted read-only on booting. For obvious reasons, this is quite annoying as most services do not start up correctly (I do not use a separate file system for /var). After the system is up, I have to log in, remount the root file system read-write, fix /etc/mtab, mount all other file systems in from /etc/fstab and then start up all the missing daemons. I know that there are ways to make a system run properly with a read-only file system, but I would rather restore the old behaviour of a writable root file system.
The strange thing is that after running mount / -o remount,rw, the file system is mounted in writable mode without any errors. I suspected some problem with fsck, but now I have disabled automatic file system checks on the partition (tune2fs -c0 -i0).When I run dmesg, only these lines mention the partition at all, although I am not sure if not something gets lost because /var/log is not writable:
EXT3-fs (sda5): mounted filesystem with writeback data mode</code> EXT3-fs (sda5): using internal journal The line in /etc/fstab looks like this:
I'm running an Acer Aspire 1830T-3721 dual-booting Windows 7 with Ubuntu 10.10 (Desktop).
Background: So first I dropped my laptop a couple feet while Windows was running. The laptop immediately shut off and then tried to boot. Booting Windows results in an unfortunate "Windows has encountered a problem communicating with a device connected to your computer. The error can be caused by ... faulty hardware ... Status: Oxc00000e9 Info: An unexpected I/O error has occurred." But Ubuntu booted fine, and could access my NTFS files fine, so I was trying to work on the problem from there. I try a few utilities, looking at the partition table, etc without actually applying any changes.
Then I run a fsck on the drive. It loudly warns me that if I continue on a mounted drive, then I'm going to mess things up. In a moment of stupidity I push on, thinking that surely it would ask me for more configuration, or confirmation, before actually starting. The fsck runs for about 1 second before I Ctrl-C it, running some preliminary stuff and then just starting pass 1.
After this, Ubuntu won't boot anymore. Instead, it hangs just after the init-bottom script runs. If I boot with init=/bin/bash, I can get to a shell, and see that my file system is still there, but not sure what else to do.
I've been running off of a SysRescCD LiveCD, from which I've looked at the drive with testdisk. Testdisk reports that "the hard disk seems too small" while showing me the partition table.
I ran a fsck on the Linux partition; it fixed a bunch of things. There has been no apparent effect on the boot behavior.
I can access all my files, back them up, and reinstall Ubuntu, but I'm hoping there's a better solution, perhaps one that will also help me repair my Windows installation (but I'm looking at one problem at a time here).
A drive on my Linux machine is NTFS as the file system. There's a file corruption issue of some kind for copying files from the drive to another or another PC result in I/O errors. Overall, I work with 2 systems, one Windoze, the other Linux. I'm about to switch the roles of the 2 machines. The one with the corrupted ntfs partition is about to become my Windows machine and the Windows machine is going to become Linux.
Since I will be installing Windows on the machine with the problematic ntfs partition, I'm figuring at some point, Windoze chdsk will kick in and fix the drive. (Windows will be installed to another drive that is perfect right now.)
Is this a correct assumption? Or, do I do everything I possibly can to fix the corrupt partition prior to the new Windows install? If this is true, what are my options for fixing corrupted files under Ubuntu? Research I've done hasn't yielded much in results and a definitive answer for fixing corrupt files in Linux.
Got Samba on fedora 13. Windows machines backup their files to the linux shared folder. I want to attach an external hard disk (USB) to the linux machine in order to backup those files. Can the external hard drive be NTFS or do I need to reformat it as Linux file system (ext3)?
I have updated my linux version 5.2 yo 5.3 after that I wanted to mount my windows drives. I installed this rpm kernel-module-ntfs-2.6.18-92.el5-2.1.27-0.rr.10.11.i686.rpm (99KB) its not working while um giving this command #mount -t ntfs /dev/sda5 /mnt shows a error unknown file system NTFS. bt it worked in 5.2.
I was follooing this instructions to repair windows system32 in this tutorial found in a previus tread, my laptop is Dell xps 2010 I had Ubuntu Live cd running with Internet,mouse and keyboard, The NTFS and NTFSProgs exist in System/ administration /synatip Package Manager,but i can mount the filesystem properly due the device name etc,thing i missing some code in terminal application.Partition table entries are not in disk order.Regard the repair of windows media center edition (windows system32 corrupted...).