I am using CentOS release 5.4 ( 2.6.18-164.9.1.el5xen ) and created an HPC cluster by using NIS ( for user authentication ) NFS ( as file system ) and mpich1 as parallel compilers and utilities and TORQUE as job scheduler. I want to make sure all users should use scheduler for job submission and should not submit the job directly ( qsub job.sh ).
I want to prevent all users from executing executable files created by self , from its home directory .
Suppose if a user create an executable a.out and if he tries to execute by ./a.out it should display an error.We should also allow users to execute normal user level linux commands .How can I implement such a set up in my environment
a small lab of linux servers contains two servers. the administrator wishes to permit user settings and project files to be available when users log in on any machine descibe the server processes needed on the servers
I got myself curious on the possibility to change user's permission to any command, or at least giving other users some command line "power". It all started a few days ago, when another user here, had a problem so that the computer wasn't answering. So, after waiting a while, chose to hit the Reset button and start the computer again. Considering this situation, I thought myself that it would have been better to restart the computer, through command line, on any terminal (F1-F6). Anyway, this user is no administrator, thus wouldn't have permission to use the code...
So, is it possible to let other users to use such commands, in order to safe rebooting the system, without logging as root?
I'm having trouble breaking down permissions in linux. Here's the scenario. I have two users: UserA & UserB with each having to ownership and access to directories myDirA and myDirB respectively.
UserA --> /source/myDirA UserB --> /source/myDirB
I need to set the permissions so that userA can access myDirA and myDirB. There are other users and directories but they should not be able to view outside of their own directories (which is the way it is now). I don't have groups set up for them and I'd rather not change anything else but just the permissions.
rwxr_x_r_x UserA rwxr_x_r_x UserB
They're read/write/exec permissions are identical.
How would i write a command that can find all the objects under the etc directory that have group write permission enabled and have not been accessed in the last X days. This is what i got from internet souce but i m not able to modify it according to my distribution. find /etc -perm -0070 -a -mtime +X ! -type l?print Here is the exact statement from link i m referring to.
I am using Fedora 14 64-bit and after login through a user other than root when I try to open any folder on other partitions I get the message that I don't have permission to access such and such folders.
I have setup a NFS server and this the content of /etc/exports at the server with IP A.B.C.D1 is:/home/shared A.B.C.D2(rw,no_root_squash)Problem is, only the root at A.B.C.D1 and A.B.C.D2 can write to that folder.
I'm a refugee from WindowsXP, running Fedora 14 with three user accountsMy problem is that I need the primary user (userd 500) to be able to have full access all other users' files in their home directories so that user can copy, move, delete, etc.I tried making that user a member of the other users' groups - but I still get the 'not got permission' error when I try to access their home directories
i have a ubuntu server and no of nodes.give me any suggesions for how to handle the node users login with my permission?.i mean than they can switching on and boot the system but the system want to stay in unusable stage.After getting my permission they can operate the system.
I've set up smbd 3.4.7 on 10.04x64 LTS server. I've set up a couple shares and I'm having problems blocking access to certain directories using native file permissions. There is one directory that has folders for each sales rep to store their current list of quoted clients, I only want sales people to be able to browse the directories owned by themselves. Everything seems to be set up correctly in terms of user groups and permissions on the filesystem.
Below is marina, a sales rep, and brian, a super user of sorts. id marina: Code: uid=1011(marina) gid=1006(office) groups=1006(office),1005(sales) id nick: Code: uid=1000(brian) gid=1006(office) groups=1006(office),118(admin),1001(full),1002(processing),1003(management),1004(it),1005(sales)
Below is the directory with all the sales reps folders. ls -la: Code: total 60 drwxrwxr-x 15 root it 4096 2011-02-10 20:06 . drwxr-x--- 9 root office 4096 2010-11-19 12:40 .. drwxrwx--- 13 katya full 4096 2010-12-07 12:36 Katya drwxrwx--- 18 lana full 4096 2011-02-08 17:09 Lana drwxrwx--- 23 marina full 4096 2011-02-10 18:09 Marina drwxrwx--- 4 mike full 4096 2011-02-01 12:42 Mike
With this setup marina only be able to browse her folder, but she can browse all folders and has full write access to all folders. This leads me to believe something is up with the smbd.conf file, which is below.
Code: [global] workgroup = COMTREAD null passwords = no server string = Root Server dns proxy = no log file = /var/log/samba/log.%m max log size = 1000 syslog = 0 .....
In this case the valid users directive would not work cause I am not making a share for each user. I had this on other shares like the db2 share. My windows box lagged heavily when I tried to access that share with an invalid user. How to deny users the ability to modify permissions I would also like to do that.
The desktop computer of my two children has a total of three users:
1) The superuser (me) 2) The user 1001 (my elder son) 3) The user 1002 (my younger son)
Both users 1001 and 1002 can not access their files system, and also they can not save any attachments from incoming mails.
What I tried so far: I accessed the file manager as superuser, and went: >Root>Home. Here I right-clicked on the folder User 1001, selected properties, selected the tab 'permissions' and allowed this user to read and write into this folder. I also checked the checkbox �extend this permission to all subfolders and its contents.
The problem is, when I reboot, everything is 'forgotten' and I am at quadrant zero again.
Eventually I should state that part of the folders are from a backup drive, because the hard disk had to be replaced so, once I re-installed the OS on the new hard drive, I copied the folders from the backup drive into the home folder.
One last question: Is there a good tutorial about permissions?
I installed proftpd on my Ubuntu 10.10 install. I also run multiple websites that I want to allow ftp access to for 2 different users. The websites are located in /home/www/. This is where the guide I was following told me to put them. I also don't have a user named www.How can I give write permission to upload, delete, and edit all the files in /home/www/ for multiple users? They can connect to the ftp server and see the file, just not change them.
I am writing a bash script to run everyday and output results to a file. When the same results are produced i want to overwrite the line from the previous day. (Or remove and add). So if the script finds a variable in a line. i want it to output the results to that line . sed -i did not work for me; sed: couldn't open temporary file ./sedTvOCEg: Permission denied
I am trying to write a script to set the SGE job execution order. I named each job with 'job1', 'job2' and 'job3'. I want my script to do: When 'job1' execution is complete, 'job2' is executed; when both 'job1' and 'job2' are complete, 'job3' is executed.
First, I tried this following script and it worked well:
I have seen command like $rm aa dd cc bb ee then something like $ >> zz // it removes zz also Where aa dd cc bb ee zz are the files in my directory. Do not know exactly.
Now my script is able to start server, But I am still have some problem with my script. When the start server command is executed, the control does not pass the line and does not execute further of that line.what is the problem and how can I get smooth execution of the my script. My Script:
I have never set-up a chroot-jailed environment before and I am afraid I need some help to do it well.To explain shortly what this is all about: I have a webserver to which users send python scripts to process various files that are stored on the server (the system is for Research purpose).Everyday a cron job starts the execution of the uploaded scripts via a command of this kind: /usr/bin/python script_file.pyAll of this is really insecure and I would like to create a jail in which I would copy the necessary files (uploaded scripts, files to process, python binary and dependencies).
I already looked at various utilities to create jails but none of them seemed up-to-date or were lacking solid documentation (ie. the links proposed in How can I run an untrusted python script)Could anyone guide me to a viable solution to my problem? like a working example of a script that creates a jail, put some files in it and executes a python script?
i need to take the backup some folders encryption formate with password i used gpg encription in the script but it is not taking the password it throos errorthe syntax is#!/bin/bashtar -cvzf test.tar.gz target ; echo "test" | gpg -c test.tar.gz --stdin
Get the list of virtual addresses being accessed by any application during its execution time. I don't want the complete memory map but just the pages which were accessed during an application's execution?
I am calling a service using http post through wget, the command is successfully executing but for each execution its creating a file and saving variable names n data n it. I want to execute this command without creation of a file. Would anyone suggest me what needs to be done in this regard.
My command: wget --post-data 'var1=99&var2=200' http://xyz.example.com:5555/invoke/Samples:httpInvoke For every execution, its creating the files with names: Samples:httpInvoke1 Samples:httpInvoke2 Samples:httpInvoke3