1- I've set up 3 virtual users,one of them is a system one (with a different password) and writes on his own home folder. With this one I haven't found any problems yet, but with the other 2 users I can't access files/folders created by them. It's a permissions problem for sure, but I'm not sure how to correct it.With these users I can upload files, create files and create folders. The problem is I can't access what I create (I can't enter a folder I created but it is there and I can upload files into it).
2- Whenever I turn on ssl_enable=YES I can't access the server (even from the server itself when I connect to localhost, It's a regular Ubuntu installation).Here's the config file for the users:
Just installed lamp, I can access phpmyadmin mysql is set up and everything. When I try to view a directory [URL] I get an access denied error.
EDIT: I changed the permissions of the "folder" folder itself, I can access everything in that directory now but not any other folders in it. Do I really need to go through every folder every time and change the permissions?
I've managed to setup Ubuntu Server 9.10, and created folders/files toview/edit/execute with Windows and Mac.
Now here is my dilemma, this is for a home server, and I will have 4 users(1 for myself, one with "admin" rights, my fiance(mac user), media pc, and a "guest" account for the computer or 2 that are out in the public(they don't need access to my taxes ).
I'd like to have it so that on my laptop, I can access the whole server, but some files/folders would even need me to put in a password. I want to do this because my fiance gets delete happy and deletes things, so if she goes on my computer she won't delete important info. I'd like her to have access to music, photos, and videos, as well as her having her own folder that she can treat as her hard drive.
The question is, do I need to setup a domain for this, or can I get away with a workgroup?
how to add files (and where) for anonymous download. I installed vsftpd and configure /etc/vsftpd.conf file...just few common options like allowing anonymous,download,upload. And now i can login with anonymous. But i dont know what to do next, i want to try to download and upload files.
I have CentOS 5.5 distribution with Dom0 and DomU installed. I try to access Dom0 files during vsftpd server from DomU during ftp client. I successfully login with root and simple user, but when I try to list (or cd to some directory) in user home the SELinux prevent it from me. I get this in audit.log:
I'd like to configure vsftpd server in a way to allow remote user (local) too see and edit configuration files in their ftp directory starting from dot (like .htaccess, for example). With default configuration + "local_allowed = yes" it does not appear to be possible:user can successfully upload .file but could neither see if it is in directory nor download it.
I've just upgraded my serv from fedora8 to fedora12 and got a problem. The httpd and vsftpd add a junk in the beginning of html files. Png, tar and other files are ok. When I make a request to localhost everything is ok.
Code: #telnet localhost 80 GET /1.html <html> </html> but from any other computer.
Code: ..... But this is not an apache problem. When I dl html files from ftp I got the same problem. Iptables is swithed off.
How yo get ride of access denied 404/403 for localhosts other folders? i am using lamp and trying to access my site which is placed in a folder wthin the document root. but its showing access denied. how i can edit permissions?
I am trying to setup vpn to access folders and files from my school. VPN is connected now through networkmanager, but I don't know how to actually gain access to the folders on the school server.
Is it through nautilus? Or the "Connect to Server"? The school provides addresses that begins with \.. something, and I have tried typing it in to nautilus and the "Connect to Server" but it fails to connect.
I have googled this but all I find is how to connect to vpn manually or through networkmanager, and not how I then can access the folders on the server.
I configured my apache2. On my Intrepid I had apache2.0 while on my Karmic I have a apache2.2. Aftere configuring I tested it and got a an error page when I tested it in my web browser. I looked into the log file that showed the following error "[client 127.0.0.1] (13)Permission denied: access to /my_dir/ denied".
It appears apache2.2 can't access directories in my home folder. File system rights for the files and folders are correct. There is no AppArmor profile for Apache. User settings in "/etc/apache2/apache2.conf" file are correct. The inaccessible folder in "/etc/apache2/sites-available/default" looks as follows:
A trick using symbolic links didn't work either. On my previous Intrepid with Apache 2.0 my pages worked like a charm. Now on my current Karmic (before apache2.conf was pre configured, now it's not) with Apache 2.2 my pages are wrecked. how I can make Apache2.2 access folders in my home folder and which settings are needed in default file for that?
I receive the message "SELinux is preventing /usr/sbin/vsftpd "net_raw" access" many times. Found this bug at redhat but really do not understand what i should do about it ((( Kindly let me know how to change this to normal. Shut down Selinux is not the way out.
I am trying to install Fedora 10 on an old P3 Inspiron 5000. When I boot from the Live cd the screen is split into three sections and is hard to navigate. I am able to edit the /etc/X11/xorg.conf file and log out and log back in to proceed with the installation. The problem is when I finish the installation and reboot to the new OS it boots to the triple section mode again because my xorg.conf is not edited. To do any commands in Fedora I need to get through a brief user setup in which I cannot navigate through until I can see the screen better. Is there a way to boot to a graphical type safe mode so I can see the whole screen to finish the installation? How can I access folders from the live cd to edit the installed files? This way I can edit the xorg.conf file before booting and all would be legible.
So, I am looking to implement an FTP server with Isolated Client accounts/directories where a client can only access what's in their directory. I also need to provide my internal user's (content managers) the ability to upload, delete, etc from all of the Client accounts. The simple part is creating the secure client accounts. It's a matter of changing DIR_MODE in adduser.conf to 700 or 770, creating a user, having the FTP server chroot them to their home directory, revoke/restrict shell/ssh access and maybe even slap on some ACL to prevent botched permissions.The hard part is figuring out how to give my power users the ability to access all of their folders without thrashing security.
My first thought was to put all of the client user-groups in a parent group and having my internal users inherit group permissions..but you can't have groups inside of groups.My second thought was to put all of the client users in the same group and prey that the FTP chroot is enough to keep them from poking around but then I have the problem of how do my internal users access other user directories if they are chrooted. Do I create a second server without chroot.do I create some weird nested homedir structure..I honestly have no idea how to satisfy both requirements (secure client accounts and privileged user accounts). I need my privileged users to authenticate against Active Directory via Likewise open, LDAP, etc and I don't care how the clients authenticate. Though, I would prefer to have both file and FTP-server level protection just to make sure no one can see the other client's data.
I just set up an VPS with ubuntu. I made a user1 and gave it ownership
Code: chown -R user1 /home/www
This user also have been given all the root privileges (I know it is not recommended!)
The problem is that each time I make new site, and user1 wants to upload (through ftp) files to /home/www/newsite I need to redo the the above command in order to be enable user1 to upload. Not only this, I need to rework permissions (744 for folders and 644 for files), otherwise the newsite throws permission errors message.
Will unzipping always append files if the directories already exist? What about tar?
I unzipped an archive the archive had 3 folders and dozens of subfolders within those. app, skin, js
The folder I unzipped it to also already had those 3 folders as well as many child folders etc.
My website is still working so I'm just wondering. Is this safe or is there anything I should worry about. The other option was to manually upload each file manually that is in those 3 folders and the dozens of sub folders.
I'm looking for a Syncronizacion Softwares (running as a daemon) to do a unidirectional syncronization of some files/folders from a lot of servers to one central server.As there are a lot of servers i want a software that only needs to be installed on the central server and pulls the files/folders from the other servers.rsync and unison need to be installed on both machines and syncronize the files/folder by sending them from al the servers to the central server(its no the central server the one who pulls the files/folder from the other servers), so they dont work for what i want.
I found that if any usual user is logged into a NDS-tree, then _local_ root has full access to user's network shares, including the user's home directory located on remote Netware-server. Is it by design or have I missed something? Nevertheless in windows local admin has no access to network resources mounted of any other user. If you runas shell (as admin) then admin in principle can't "see" network shares which were mounted (connected) by other users - they are accessible ("visible") per session.
I'm new to Linux and have taken over at a company where the linux admin walked out. he has several older linux based servers running version 8.1 i tried running the command sudo apt-get install swat xinetd to setup swat but i get an error that it can't find the package. in addition, this directory xinetd is no where to be found under the etc directory. i've verified that samba is in fact installed on these servers but for some reason the xinetd and inetd directories are not on these servers.
Could it be possible that this older version of linux does not support Swat? i'm considering upgrading very soon to the latest but my main problem right now is i need to copy all web server html files that he had on the web Server i know there are different ways to copy files from linux systems but need the best way to copy all the files so none of the website links get broke off this linux server over to my windows server.
Since i don't have access to anything of the directories nor files on the web server, what is the best way to copy these files and folders off this server?
I have a server setup with all my web development stuff in /var/www and in several sub-folders within that. (each project having it's own folder)It works great with one FTP account. But recently I've been getting help on a projects from a buddy of mine that freelances, and have made him an FTP user account as well. All is fine, except for when he tries to edit a file and gets a permissions error.
Here's the issue, I don't want us to have the same FTP login, but all the files are currently owned by my user name. So, when he logs in to edit a file, he can't because I'm the owner, and the files are set to 744. Will I cause any harm by adding both users to the same group (www-data) and chmod'ing the files to 775 so that we can both access and modify the files?
I am currently new to all of this and not even sure if I am in the right place.But I am currently running a server with OpenSSH.When you go to my site's homepage, I have a link that redirects you to my files that I will be sharing.They are currently access denied as I do not want the public to be able to download.My question is how would I go about having a login box to pop up when they clink on the link on my homepage to access the files folder.Currently I have accounts set up and it works when going through Filezilla with SFTP but I'm not sure how to give access to use via a web browser