Ubuntu :: Multi Users Accessing A Remote Server On Local Network
Jul 24, 2010
I would like to experiment a "green" idea of virtual desktop where multiple users are served by a single powerful machine.
I have a server running 24/7. The monitor of this machine is turned off most of the time and the OS is on the login screen.
Other users, in the same local network, use less powerful machines, which could be a thin client or an old Pentium 3 machine. They access their accounts remotely and work with the GUI as if they were sitting in front of the server. Each user sees their own desktop (different themes, screen resolution, etc.). And of course it can happen that several users could log in at the same time.
The usage is modest: mostly web browsing and the usual default applications (office, wine, gimp, etc.). In particular no games or any demanding applications. The users want to use their desktop in graphical mode only.
Question: How do we call this way of using a server? Is it possible with Ubuntu? And how to implement it?
I am having problem to collecting email from remote POP3 (all the emails for a domain is stored here) and distribute it after collecting to several users defined to Linux server. I have install postfix in Linux server for email distribution.
I run a system that users may log into either remotely or physically. Multiple users may be logged in simultaneously because of the remote access, but only one user can be physically logged in at a time.With the current setup, however, if the physical user inserts a flash drive (which the OS mounts automatically) then the remote users gain access to the removable media.
i have php5 and mysql on ubuntu server which is hosting a WordPress blog. i activated it remotely over the internet and it works fine still over the internet. now im home and using my local network again and now when i try to go into it with its local IP, it just shows some text (blog title, posts and so on) with NO pictures.
I've just added my wife as a seperate user on my desktop and have a question about shared network folders. So /etc/fstab mounts network folders from a second computer and until today I've mounted them to /home/David/NetworkData
This of course means that when my wife logs in she won't see them since they're not mounted to her home folder. So what folder should I use and what tricks so that we both have it visible and accessible in Places from the top menu?
I just installed open Suse 11.3, and I cannot SSH my school. Upon further investigation I could not even ping any machines outside my local area network. Ironically I could nmap machines outside my local area network.
I've got a lamp solution deployed that I didn't write but I do have root access to the server. What might be the best way to determine the number of concurrent users accessing this web app throughout the day?
I have a home network set up that includes windows xp and a local server. DNS resolution allows both machines to access the internet. I know this because firefox works on the xp machine, and yum works on the fedora machine.I can see the server index page through firefox by ip. I can't see it by url.
I cannot access a remote server(Ubuntu) using ssh or even using sftp from my ftp client. I've been accessing this server regularly for many months with no problem until now. There is a web server running and that's OK, so no problems there. I asked the owner of the server if he could access using ssh from within his LAN, which he could! So openssl is working from within the LAN, but not ouside of the LAN. I access the remote server by using one of [URL] free domain names. I'm aware that this service requires renewal or activity at least once a month and as far as I know this is up to date and working OK.
Is this a router gateway problem? I suspect that port 22 has been blocked, but what could block this port? It's been working well for so many months. As far as I can tell the owner of the server/LAN has not been tampering with the router, so what could be blocking port22? I'm currently accessing the server using vnc4server.
Another really weird thing is that when I access the terminal on the remote server via vnc4server and I ssh back to the server using one of [URL] free domain names, it works! So as I understand ssh is accessing itself by going out of the LAN and back to itself. This would suggest that port22 on the gateway router is working. But when I try to ssh from a pc outside of the LAN, I just get timed out!
* openssl working from within the LAN but not outside of the LAN. * sftp not working. * domain name from dyndns.com working and up to date. * Web server is working OK.
i have made a java web server which works on localhost.but now i want to capable it handling many clients at a time.so clients running on different computer need machine name or IP address of server computer.How can i do this in java?
I was installed xming application on windows XP & Windows 7to access Linux server's GUI session and i was able to access it using root account but when ever iam trying to access GUI using normal account it is giving some error.I am testing this since it is a freeware.
Which is the best (easy to follow the chat not like wireshark) to sniff msn chat for remote users in my network? I was try msn shadow that it has easy to understand inderface but it sniffs only local chat.
I've setup a Lamp Server for Testing, The Lamp Server is Up & Running on CentOs 5.5
I am now trying to setup a VSFTP server where local users can upload files to there home directory so that Apache can serve web pages straight from the directories of system user home/accounts giving users the ability to run their own web sites which are hosted off the main server [tutorial here: [url]
So far i have been able to serve/display index.html files from the users home directory [url] but so far i cant upload files to any user home directory, every time i try to upload a file with filezilla i get this error message: 553 Could not create file. Critical file transfer error
I have searched online for similar problems like mine and so far i've tried alot of the solution but none seem to work. I'm confused, dont know where i went wrong, i put the users in a group called ftpusers and here are the permissions on the users (test, ftpuser & testftp) home directory. have a look an tell me where i went wrong :(
Also the root directory where the web pages are served from is called public_html here are the permissions
Here is my vsftp.conf file can someone check it to see if i made any errors in there:
Running Ubuntu 9.10. In the Remote Desktop config dialog I get: "Your desktop is only reachable over the local network. Others can access your computer using the address 127.0.0.1 or tabatha.local." I understand this means only the loopback ip address is available. All my other machines show their true local ip address (e.g., 192.168.1.104) in this dialog. Thus I cannot log on to this desktop from other machines.
When I try to do a remote logon from another Ubuntu 9.10 box (or from an XP box using a VNC viewer), I get: "Connection to 192.168.1.102 has been closed." What steps are needed to make this machine show its actual ip address? All file sharing between the various machines is working properly and all windows shares back and forth between XP and 'nix, and among the the vaious XP boxes and linux boxes are available as designed.
The issue is that my CentOS workstation is in a vlan from where the Intranet's DNS servers are unreachable. For browsing the web there is an ISA proxy server, which I presume resolves DNS for my firefox. However, wget, host, ping and aria2c fail to get any sort of DNS resolution since they're being run from command line.I have exported HTTP_PROXY value, which provides me internet access on console, but,only when I connect using IP address. It fails on name resolution.
My question is:May I redirect the DNS queries to my home PC which would be running a DNS server on a non standard port?I was thinking of putting nameserver 127.0.0.1 in /etc/resolv.conf and then put iptables rule to redirect 127.0.0.1:53 UDP to a.public.ip.address:3535 UDP..I don't know if I am shooting blanks or what, I am not very much aware of this kind of setup.My main need is to provide DNS resolution to console apps.I want to utilize my company's idle bandwidth for bulk downloads, so, using proxy, SSH tunneling through my Home PC is out of question.
I was looking for:how to access my Linux machine(OpenSuse 11.3) that is being hosted at my hoster company from my home Win Pc, I found TightVNC , but I am confused, should I install it on both systems ( viewer and server) ? .. what about the VirtualBox , can i Access with that tool or it is only for mounting the local Virtual Machines?
"Merging" may not be quite the right word but that is the desired end result.
Scenario: many Solaris 10 servers, each with various local users. We want to set up LDAP for all for all of them. LDAP server is set up, procedure for getting other servers to use it for user authentication is documented and tested. The question is how to handle users that are in LDAP who also exist as a local user on a given machine.
It appears that the usernames on both sides follow a convention and therefore match but obviously the userids will not match. Local user joe has userid 1234, LDAP user joe has userid 56789.
The way I see it we'll have to:
1. move local user joe's home directory to the path that LDAP user joe will want
2. change local user joe's userid to that of LDAP user joe
3. change joe's files' owner to his new userid
4. remove local user joe
5. finally configure LDAP
Is this a rational procedure? Is there a more effective method? I'm not looking forward to this as there are many servers and each of them have a different set of local users, each with different userids which will have to be handled manually and individually therefore not even scriptable much.
I have a remote drive mounted on my system(ubuntu 10.04 x64), and i have the contents of that drive backed up to dropbox. the problem is, if i unmount the drive, the files disappear from dropbox. is there a way to mirror the contents of the network drive to a localfolder(preferably in such a way that all changes and file deletions are changed on the local folder instantly, but unmounting doesn't delete it all)? It looks like rsync would work, but im not sure how to make it work.
I installed sendmail in RHEL5.4 with TrendMicro Spam scan engine. The configuration like sendmail should forward all the mails to scan engine after scanning it will deliver it to the mail domains. same way all the mails coming from external servers are scanned and the delivered to local box. My problem is when i sent mail for local users its delivered locally. but when i sending mail for external like yahoo its going through scan engine. I added the smarthost in sendmail.mc file also.
As many developers probably do, I have a Windows based machine on which I run XAMPP locally to test my code and a Linux machine with Fedora as my remote server.As I sometimes use .htaccess as a way to authenticate some parts of the website, I end up having two .htaccess files: one with the local path (something line D:My_Webs) and one with my remote path (something like /var/www/html/) to the password file.I have searched high and low, but I cannot seem to find any trick so that I only have to maintain one version of the .htaccess file which can work on both Linux and on Windows machines.
We are using postfix for relaying our internal system generated mail to other domain through google apps. All client servers able to send mail through the postfix server, only user of local postfix server(root, etc) unable to send mail with command like mutt, mail.My detail configuration given below: