Networking :: NFS Files On Client Have Different Owner/group Than On Server?
Feb 4, 2011
The server is named alpha and is running Archlinux. It is exporting a directory named /files. The server is a couple of years old and I have accessed it extensively from clients running Arch, Suse, PCLinuxOS, and maybe some others, all with no problems. The clients (3 of them) are new installations of Linux Mint 10 (Julia). When I mount the nfs all of the nfs files are visible as expected. However, the owner/group is drastically different than on the server.
I might add that I have set up user id's and group id's the same. My user is 1003 on all systems, and the users group is 100 on all systems. When I am on alpha (via ssh), here is a partial file listing.
[dick@alpha dick]$ ls -l
drwxr-xr-x 3 dick users 4096 May 16 2009 airplane
-rw-rw-r-- 1 dick users 240978 Jun 27 2009 Alice Grad 1934.pdf
-rwxr-xr-x 1 dick users 444 Jul 8 2007 alpha2ast
-rw-r--r-- 1 dick users 444 Sep 2 2009 alpha2charlie
If have searched the Mint forums, LQ forums, and google in general. I must be missing something in my search because I can't believe that no one else has this same problem and I am having it on 3 different boxes.
I can't figure out how to make files have a different default owner:group.. Example:I need the users of my group called gpib, to create new files with: username:gpib, instead of the default: username:username
have recently installed ubuntu server on a new machine. I have added 3 users and I have assigned them to a group.The three of us work together on a lot of stuff so what I would like to do is to have a specific folder made the groups folder. All files that are created or moved into this folder should automatically be owned by the group. I.e. all 3 of us should have the right to read and write to these files.
On the server I can see the owner as vampird and the correct group, VampirD Microsoft Windows is like air conditioning Stops working when you open a window. -----BEGIN PGP SIGNATURE----- Version: GnuPG v2.0.12 (GNU/Linux) Comment: Using GnuPG with SUSE - [URL] iEYEARECAAYFAkxHNfAACgkQJQ+0ABWtaVlcagCdEo5kiwydUTmZ+dkD4R4jholx bi4AoO6T2OzHealqsQ+9Z42jJ7rYJ6uL =YKm8 -----END PGP SIGNATURE-----
In Windows I used WinSCP to do all of my server work. It was easy and intuitive to use. In Ubuntu, I've been recommended to use "sftp://" for the location. I can change folder permission settings this way, but it doesn't allow me to change the owner and group, and doesn't allow to change the file permissions (folder permissions are ok though).
Can anyone point me in the right direction? How do I go about doing this? I'm much more comfortable doing it via GUI rather than terminal.
A colleague of mine has a Linux box (running Debian I believe) with an SVN repository on it. The repository directory and files 'owner' is my colleauge. We are both members of a group called 'users'. He manages several projects both Linux and Windows apps, while I have one Windows app. For the Windows apps, we both use TortoiseSVN via an SSH link to commit/update. Performing the command 'ls -l' shows the repository files and folders on the Linux box to have the following permissions:
-rwxrwx--- john users
However, when my colleauge commits to the repository, the permissions change to:
-rwxrwx--- john john
This then means I get 'Permission denied' when trying to access the repository myself as it appears that the group permissions have been overwritten with only 'owner' permissions. To fix this, a 'chown -R' command is applied to the files/folders to set the permissions back to owner/group, but each time he writes to the repository, the issue repeats.
I am creating a tar gzipp'ed archive on my local machine (as user1) using the following commands:
user1@devmachine:~/$ tar czpf dir.tar.gz thedirectory on the server, I untar it (as user 2) using the command user2@servermachine:~/$ tar xzpf dir.tar.gz
I find that the extracted files are owned by another user (say user3) What is the logic that is used to determine file ownership if the owner of the extracted file is not a user on the target machine? I am running Ubuntu 10.0.4 on both machines
I use 9.10 desktop with a root user and my own user (timmo), I did not create anything else. Now I check a directory (mysql databases) with ls -l and I see mysql not only as a group but also as owner. How can mysql, not being a user on my system, be an owner? In users and groups I see that all of the many groups only have two members, root and timmo. I know that mysql users and linux users are different animals but ls -l is definitely a linux command.
Did a fresh install of Maverick, all is well but if I insert a video DVD, it won't play. But if I open Movie Player, etc. as root, I can play the individual chapters - that is I need to manually choose which chapter to play, it won't start at the beginning and play to the end. Have installed libdvdcss3, restricted extras, etc. I am a member of the "video" group.Data & music CD's work fine in the drive; data DVD's are fine also. Just no DVD playing with me as the user; nor does the DVD appear in my Places menu, etc
According to a couple of different places, it's not possible for me to put a line in /etc/fstab to mount one of my partitions with owner and group not root; instead, I have to mount it in /etc/fstab, then chown & chgrp to my user. That seems ridiculously tedious and silly... is it true? I'm sure a short script could be written to get around it, but it seems obtuse for Linux not to allow that to be set in /etc/fstab.
In my /var/www directory, I have everything set up with: user: www-data group: developers directories: chmod 570 files: chmod 460
Everything seems fine. Users from the developers group can edit files and all, but now we began using the Git repository, and whenever a user edits a file (ie. Joe who is a developer,) file permissions get screwed again. Now they're: user: Joe group: Joe directories: chmod 755 files: chmod 644 How can I fix this so permissions remain the same?
I'm carrying out a project for my university (CIT in Cork, Ireland) and I'm using CentOS running over WMware. I have a server and a client. The server has no GUI (command line UI) while the client has a UI. I need to install a Simple Forum Machine application and I'm told to FTP the files into the server. I figured out that the best option is to load the files in the client via the GUI and then ftp them in the server. How do I transfer the files from a the client t o the server using FTP? I'm totally new to Linux so the more details the better. Also I'm trying to mount a USB key on the server but have had no luck.
i want secondary users can able to change the files permissions of primary group?user MAC is having www as a primary and httpd as secondary group. But he want to change the file permissions (chmod) httpd group files. Is it possible or not? I think its not possible. If it`s possible then let me know how?
I have a disk that I access from several locations. I require that all the files created always have the same owner, at least group wise. I have different users (that are in that group) that need to be able to read/write all these files. I have these users access the files over samba, and sometimes locally on the server. I know that you can do something with a thing called sticky group or whatever, that files created in a dir with that flag will get the same group, but it has not worked consistently so far. It must also work for directories created by these users.The permissions should be 770 (chmod).Is there a way to set this up, that all files created always have the same group? Right now I am running a cron job every hour or so, to chmod and chown all the files to the right group, but this is far from elegant of course
On my RHES4 I noticed a load of files which had owner set as the owners uid rather than the actual username - is this usual behaviour ? On a similar system the same files actually have the username as the owner.It's just causing me issues as I have changed the users ID and now some thing's wont start meaning I have to manually do a find and chown on the system.
For example /dev/loop*, /dev/raw/*, etc., they are automatically reset to root/root after rebooted.Change the owner/permission of device files maybe not a good idea, though. I just want to know if it is possible and how?
I'm trying to understand the last few hours... I installed slackware 13 yesterday in a multiboot system. On a seperate hdd from all the OS's I have my mp3 collection......I could play the mp3's as root after manually mounting sdb, but as a user I was unable to play them even though I chown'ed and chmod'ed 777 until I mounted sdb in fstab. The second drive was formatted ntfs by vista.
For some time now, I'm having some problems with configuring an NFSv4 server to let it work with a firewall. I've already searched to web, but I was unable to find a solution that works for me.
The situation is as follows: I'm trying to connect an NFS client to an NFS server that is behind a firewall. I don't have access to this firewall, but I can contact the administrator to open some ports for me. I already did this for opening port 2049.
The result is that the client can read files from the server, but is unable to write files to the server. I believe that for writing an extra RPC-connection needs to be set up. However, the ports on which the RPC-connection is set up, seem to be different for every connection (I verified this using 'netstat -tn').
Clearly, this is a problem since the server is protected by the firewall.
Thus, what I want to do is configure the server in such a way, that it always uses the same server-side port(s) to connect with the writing clients (just like 2049 for reading). I've already tried to configure the /etc/default/nfs-kernel-server and /etc/default/nfs-common files, but that hasn't really worked out yet.
Note: Because I don't like to contact the system admin every day, I hooked up 2 computers (client/server) on which I set up the same configuration (without the firewall). I'd like to see it working on those machines first (that is, 'netstat -tn' showing the correct port), before I contact the admin to open some extra ports.
I just finished set up a PXE server, and was wondering is it possible to automated save/redirect files from the client into server? Basically, test will be running on client, and wanted to save the result(text file) into server after done testing.
I can see the owner and group ids are shown because there are no corresponding entries in /etc/passwd and /etc/group respectively. I don't know much about linux and dare not to edit these files, I wonder if somebody already knows whether linux would map the owner id of files coming from other computers to the account name in /etc/passwd and display them when necessary (for example, when using ls -al)?
CentOS 5.4 install, likewise open standard install (For active directory authentication).I have a license service which requires a license.txt be in the users home directory.The group owner for license.txt must be the same as the license service. Whenever a new domain user logs in, it creates the all the appropriate files but the group owner for license.txt is the users domain group. My current workaround seems like more effort than it's worth, is there another way to get this process solved easier/more secure?
- copy the license.txt into /etc/skel
- created a script to check for the presence of license.txt, check it's permissions and change them if necessary
- gave the domain's group sudo [nopasswd] access to the script (the script is not writable)