Red Hat :: Files Have Owner As Uid Number Rather Than Username?
Nov 18, 2009
On my RHES4 I noticed a load of files which had owner set as the owners uid rather than the actual username - is this usual behaviour ? On a similar system the same files actually have the username as the owner.It's just causing me issues as I have changed the users ID and now some thing's wont start meaning I have to manually do a find and chown on the system.
How can I make a virtual host (right now I just use NameVirtualHost *:80) that will load the same page for every domain that matches imap.domain.com, smtp.domain.com, or pop3.domain.com?
I have a disk that I access from several locations. I require that all the files created always have the same owner, at least group wise. I have different users (that are in that group) that need to be able to read/write all these files. I have these users access the files over samba, and sometimes locally on the server. I know that you can do something with a thing called sticky group or whatever, that files created in a dir with that flag will get the same group, but it has not worked consistently so far. It must also work for directories created by these users.The permissions should be 770 (chmod).Is there a way to set this up, that all files created always have the same group? Right now I am running a cron job every hour or so, to chmod and chown all the files to the right group, but this is far from elegant of course
I am creating a tar gzipp'ed archive on my local machine (as user1) using the following commands:
user1@devmachine:~/$ tar czpf dir.tar.gz thedirectory on the server, I untar it (as user 2) using the command user2@servermachine:~/$ tar xzpf dir.tar.gz
I find that the extracted files are owned by another user (say user3) What is the logic that is used to determine file ownership if the owner of the extracted file is not a user on the target machine? I am running Ubuntu 10.0.4 on both machines
For example /dev/loop*, /dev/raw/*, etc., they are automatically reset to root/root after rebooted.Change the owner/permission of device files maybe not a good idea, though. I just want to know if it is possible and how?
I can't figure out how to make files have a different default owner:group.. Example:I need the users of my group called gpib, to create new files with: username:gpib, instead of the default: username:username
I have a directory cookie_tmp which is owned by some:fella. Session cookies are being created under this directory as How can I set the directory so that files are created and owned by some:fella ?
The server is named alpha and is running Archlinux. It is exporting a directory named /files. The server is a couple of years old and I have accessed it extensively from clients running Arch, Suse, PCLinuxOS, and maybe some others, all with no problems. The clients (3 of them) are new installations of Linux Mint 10 (Julia). When I mount the nfs all of the nfs files are visible as expected. However, the owner/group is drastically different than on the server.
I might add that I have set up user id's and group id's the same. My user is 1003 on all systems, and the users group is 100 on all systems. When I am on alpha (via ssh), here is a partial file listing.
Code: [dick@alpha dick]$ ls -l total 9740 drwxr-xr-x 3 dick users 4096 May 16 2009 airplane -rw-rw-r-- 1 dick users 240978 Jun 27 2009 Alice Grad 1934.pdf -rwxr-xr-x 1 dick users 444 Jul 8 2007 alpha2ast -rw-r--r-- 1 dick users 444 Sep 2 2009 alpha2charlie
[Code]...
If have searched the Mint forums, LQ forums, and google in general. I must be missing something in my search because I can't believe that no one else has this same problem and I am having it on 3 different boxes.
have recently installed ubuntu server on a new machine. I have added 3 users and I have assigned them to a group.The three of us work together on a lot of stuff so what I would like to do is to have a specific folder made the groups folder. All files that are created or moved into this folder should automatically be owned by the group. I.e. all 3 of us should have the right to read and write to these files.
I'm trying to understand the last few hours... I installed slackware 13 yesterday in a multiboot system. On a seperate hdd from all the OS's I have my mp3 collection......I could play the mp3's as root after manually mounting sdb, but as a user I was unable to play them even though I chown'ed and chmod'ed 777 until I mounted sdb in fstab. The second drive was formatted ntfs by vista.
I can see the owner and group ids are shown because there are no corresponding entries in /etc/passwd and /etc/group respectively. I don't know much about linux and dare not to edit these files, I wonder if somebody already knows whether linux would map the owner id of files coming from other computers to the account name in /etc/passwd and display them when necessary (for example, when using ls -al)?
CentOS 5.4 install, likewise open standard install (For active directory authentication).I have a license service which requires a license.txt be in the users home directory.The group owner for license.txt must be the same as the license service. Whenever a new domain user logs in, it creates the all the appropriate files but the group owner for license.txt is the users domain group. My current workaround seems like more effort than it's worth, is there another way to get this process solved easier/more secure?
- copy the license.txt into /etc/skel
- created a script to check for the presence of license.txt, check it's permissions and change them if necessary
- gave the domain's group sudo [nopasswd] access to the script (the script is not writable)
I am using the diff command with the -r option, to compare a large number of files and files in subdirectories. My main interest is to find out which files have been changed, and not what the actual changes are, and since a lot of files has been changed, it would be a lot easier to view the file names only. Is there and option for diff that might do this, or does there exist a similar tool/command that could do the job?
I installed Fedora 12 a couple of days ago. This is my secondary operating system. The main one is Mint 8 and in order to have the same documents in both OSs I have a separate /home partition, but for some reason I don't see any of my Documents. In the installation I told fedora that I already had a /home partition and to use it. Also if I mount Mint's disk, while I can browse the folders, I can't see any document .
Finally I can access my Mint's home folder and files but since my username in fedora has a capital A it created another home folder. How can I change my home folder path so it point to the other home folder?
15 this is a sentence containing various words and spaces 34 this is a another sentence containing various words and spaces
cat file2.txt
2 this is sentence1file2 6 this is sentence2file2 54 this is sentence3file2
I would like to join these 2 files. The result should look as follows :
cat joinedfile.txt
2 this is sentence1file2 6 this is sentence2file2 15 this is a sentence containing various words and spaces 34 this is a another sentence containing various words and spaces 54 this is sentence3file2
==> so the joined file must be sorted on the first number. Any ideas how this can be achieved ?
Is there a limit to the number of files ext3 can support?
Reason I'm asking is because on one of my internal drives, I have around 750,000 files. The drive is 500Gb and currently using 150Gb... I noticed recently that when I try to copy a new directory or file, the transfer rate is extremely slow at times. It is sataII and sometimes it gets as low as 500kb/s (yes, kb!)
Would somebody please shed some light?
I noticed it might be related to the process gvfsd-metadata
I was nosing around in my /home folder and I noticed that the /.thumbnails directory had 38,000+ files in it. That number seem a bit excessive to me. Is there a way to limit the number of files that are allowed to be in that directory, and maybe delete the oldest files automatically when the directory reaches it's limit in order to make room for the new incoming files, so there are no "directory full" type of errors?
Getting together a script that will add numbers to all the files in a folder.
I've ripped most of my CDs to oggs for my new pmp, but I found that the pmp doesn't like files that are numbered just as 1 and 2, as it thinks that the 2 is more than 10.
So instead of going through all of my music folders and renaming every file by hand from 1 to 01 and from 2 to 02, I'd ask if there's a script that can be executed to add these numbers for me. It'd be even better if it only added the number to the files with only one digit.
Here's an example:
I want to rename:
And I'd like to do it to all single-digit files lower than 10 in the folder, if possible. If not, I can isolate them by hand.
I would like to ask you if there is any maximum allowed number of files per folder in linux (without risking it to lose everything). I am using openuse 11.4 with latest kde (4.6?).
I am trying something fast and dirty and it might be that one folder will contain like 10^6 files.
Is there is anything I should be warned about that?
i need to know how to find number of files in a directory? is there any system calls in fedora 12.And i need to know how to perform a operation if the that count increases by one?
Does anyone have a solution for merging files if the number of rows in the two (or more) files is non-equivalent.To exemplify, how about merging the following 3 files:
I currently have a problem in running rsync on 64 bit Debian Jessie (although the problem also occurred with 64 bit Debian Wheezy)I am trying to use rsync to archive my home directory (which is on a hard disk) to a USB memory stick. The home directory is about 18GB in size and the memory stick has 32GB.
Unfortunately, rsync hangs after copying a certain number of files and the process eventually has to be killed. Rsync was rerun but hung again at about the same point as before.This has now happened several times. Each time the hang occurs at about the same point.Use of strace after the hang shows that rsync appears to be processing a pdf file at the time, although not always the same pdf file.I originally had the rsync hang problem on a PC which ran 64 bit Wheezy and which used a USB 2.0 port.
I now am running rsync on another PC, which runs 64 bit Jessie and which uses a USB 3.0 port..I have also tried three different USB sticks, two from one manufacturer and the third from another manufacturer.All give similar rsync hangs.