My university allows me to connect remotely to my storage space. In Windows, I can use: Code: \name.of.server.ca and after I type my userID and password I can map the folder to a drive letter (say, X:) In Fedora, I can connect remotely with Connect to Server using sftp/ssh (although I've not been able to store the userID and password temporarily). What I'd like is to mount or map the remote folder in a way that would allow me to access it from applications (for example, to call a files on the server from within LaTeX ).
I would like to read and write to a folder in an Ubuntu VPS remotely via a Windows Explorer (Windows 7 Ultimate). Considering that my VPS has very little resources (128MB RAM), what program should I use?
I want to mount a remote drive on bootup. I'm using FC14 and remote machine is FreeBSD. I've written a shell script to mount it. The script contains only one line: Code: mount 192.168.1.33:/home/user7 /media/mc33
I've to run this script from superuser mode to mount the file system(it works). So to mount it at bootup, I added the shell script to my PATH (/home/me/bin) and added it to Menu->System->Preferences->Startup Applications. Well this doesn't work because the root privileges are not present. I tried fixing it by giving root privileges to my shell script Code: #chmod +s mount-mc33 but it made no difference.
My linux PC is connected at work and I was able to configure the IP to connect to the network. I need a way to access the linux computer from other windows PC in the office via shared folders and or remote desktop. I am stumbling upon "VNC" on google.
I am having trouble mounting a nfs network partition. I think my problem is related to the firewall in my server.
1. - When the firewall is disable in the server, all the clients are able to mount the network partition without a problem. This is good, but I want the firewall active.
2.- When the firewall enabled, without any port opened, none of the clients are able to connect to the server (expected behavior). This is the message I got:
mount -t nfs 192.168.1.103:/home/username /mnt/myHomeCaraota/ mount: mount to NFS server '192.168.1.103' failed: System Error: No route to host.
3.- When I open port 2049 only (see lines below) Nothing changed. I got the same message. No connetion
-A INPUT -m state --state NEW -m tcp -p tcp --dport 2049 -j ACCEPT -A INPUT -m state --state NEW -m udp -p udp --dport 2049 -j ACCEPT
4.- If I open ,additionally, ports tcp/udp 111 thre is an improvement because at least I get:
usename@laptop ~]$ sudo mount -t nfs 192.168.1.103:/home/username /mnt/myHomeCaraota/ mount: mount to NFS server '192.168.1.103' failed: timed out (retrying). mount: mount to NFS server '192.168.1.103' failed: timed out (retrying). mount: mount to NFS server '192.168.1.103' failed: timed out (retrying).
[Code]...
But in the last case I am opening everything to the clients (192.168.1.100-192.168.1.104) and I believe it is not a good use of the firewall. In summary, I think my nfs configuration is o.k, because I am able to mount the partitions when the firewall is down or when all the ports are open for a given number of clients. I also believe that I need to open ports 111 and 2049 because at leas the server is responding, but I know that I need something else, but I can not figure out what that is.
I have created a folder including some html link pages in /test/htlink . Now I want to create a soft link of that folder as htlink -> /test/htlink in /var/www/html . Now when I m browsing the folder , its erroring (404 Not Found) folder is not found on the server . I have given 777 permission on every file-folder in /test . No firewall & no SELINUX .
I setup a Samba share and I cannot connect. I can mount in on local host but when I CD to the folder I mounted the share on I get access denied when I run ls.
Running NFS on Fedora 10. Exports fine. I tested it locally. I tested the NFS configuration by trying to access the exported directory from my local machine, before testing it from a remote machine. While logged in as root, I created a new directory "/mnt/nfstest".
Then I mounted the NFS share at the new directory I created: [root@eric root]# mount -t nfs localhost:/mnt/nfs /mnt/nfstest
When I tried to mount on the remote client: [root@frank root]# mount -t nfs eric:/mnt/nfs /mnt/nfstest
After a while I got: # mount eric:/mnt/nfs /mnt/nfstest mount.nfs: mount system call failed
I tried strace but wasn't sure what I was looking for, but I've attached the results as a .odt file.
I can not use nfs from F10 client to F12 server. nfs mount on F10 to F12 times out anf nfs4 mount gives "mount.nfs4: mounting localhost:/home failed, reason given by server: No such file or directory" I have tried to close firewall and set selinux to permissive mode on both client and server with same result. Samba works fine. On server [root@flokipal ~]# mount -t nfs4 localhost:/home /media/tonlist mount.nfs4: mounting localhost:/home failed, reason given by server: No such file or directory
but
[root@flokipal ~]# mount -t nfs localhost:/home /media/tonlist [root@flokipal ~]#
I recently configured a server that I would like to access remotely. I want to start x through ssh and then log in through vnc, but I get prompted to type in my password in gnome before I have a change to login. Which is a problem because you need to be at the box to enter the password. Is there a way to authenticate this through ssh or give vnc automatic keychain access?
I am trying to mount dosbox on a share on a remote computer. I did it in Win XP without any troubles. In Ubuntu it does n't work.
First I made the map in media and connected the share. The share works fine, but I cann't mount DOSBOX. Then I put the share in the /home/user directory. Share works fine! But dosbox does not mount.
I have Fedora installed on a netbook. I customarily mount several NFS shares on this machine, from both a desktop system running F11 and a small server running FreeBSD. On the server side the shares are write-enabled. On the server side the shares are write-enabled. In the past this has worked fine.
However since upgrading to F12, my configuration no longer works as before. Reading from the NFS shares is no problem, but as soon as I try to write to one, either in Nautilus or from any other program, including on the cmd line, all hell breaks lose. Nautilus crashes, and I am unable to remount the shares. Usually rebooting the client is the my only recourse.
There are no clues in dmesg on either the client or the server. In terminal trying to remount a "trashed" share I see this:
Code: $ sudo mount -v venus:/media/disk8 /media./disk8 mount: no type was given - I'll assume nfs because of the colon mount.nfs: timeout set for Sat Feb 20 17:34:59 2010 mount.nfs: prog 100003, trying vers=3, prot=6
[Code].....
The NFS versions under F11, F12, and FreeBSD are the current ones (all updates applied).
Attempting to create a backup script to copy files from one file system to a remote file system.
When I try this I get:
Quote:
# tar -cf - /mnt/raid_md1 | gzip -c | ssh -i ~/.ssh/key -l user@192.168.1.1 "cat > /mnt/backup/fileserver.md1.tar.gz" tar: Removing leading `/' from member names Pseudo-terminal will not be allocated because stdin is not a terminal. ssh: Could not resolve hostname cat > /mnt/backup/fileserver.md1.tar.gz: Name or service not known
[Code].....
I know that the remote file system dir is RW and the access is working fine. I am stumped...
I have two remote systems which I have fstab entries for on the local system. I have them set to noauto, because mount fails during boot for some reason, but that's not the problem. For years I've mounted them in rc.local.
The problem is after I recently reinstalled Debian, when I mount them manually it always asks for my user's password. I've copied my user's pub key from the local to the remote system and put it in the user authorized_keys file (not root), like I always have. But it still asks for a password, and so fails to mount in rc.local.
I have a machine running Ubuntu 10.04. I have a VPS (virtual private server) also running Ubuntu Linux. The VPS has a dedicated IP.
I want to have a setup, so that the directories on the VPS are accessible on my HOME PC like a mounted drive.
That is I should be able to do things like edit files on the VPS and copy/move files from/to my home PC to the VPS, as if the VPS was mounted on my box.
I have a computer (C1) to which I can connect through the Internet (ssh, for instance) (it has a static ip and though it is sitting behind a router, the appropriate ports are all forwarded). I have another computer (C2) that doesn't have a fixed ip address and sits behind a router that I cannot fiddle with (so no port forwarding here). I would like to know if there is any way to connect from C2 to C1 such that a directory on C2 would be mounted on C1.
From a Fedora 9 server (I think!) I'm trying to mount an nfs drive from a Mac Server. You see, a couple weeks back this was working fine but now there seems to be a problem. I've tried using the mount command and found the following line in the /etc/fstab doc...
Code: <IP ADDRESS>:/remote/nfs/directory /directory/to/mount/on nfs defaults 0 0 It was commented so I uncommented it, hoping this would be a quick and easy resolution =). Anyway, it didn't work and when I run mount -av the server timesout. So on the Mac Server, I enabled Samba sharing preferences and tried adding this to fstab,
Code: <IP ADDRESS>:/remote/nfs/directory /directory/to/mount/on cifs username=name,password=password 0 0 When I run mount -a it hangs. I guess I'm doing everything totally wrong. I can SSH into the MAC Server and use tar, rsync etc with no problem. But I really need this drive to mount as it once use to. It's been suggested to me that the NFS or networking module on the Linux box might be broken.
I'm setting up my fedora 10 server. I am currently working on samba. I have added users and created their home directories with their own html directories. I have changed their html directories to 777. However, it appears to be read-only since I cannot create files or folders within it. Can anyone please give me some ideas as to what may be causing this? Probably some extra configurations I may have overlooked? Any help would be kindly appreciated. I also created a general wshare (writable share) directory for all users within home directory but I am encountering the same problem and cannot write to it.
I've run into some problems trying to change the default cgi folder from /var/www/cgi-bin to /var/www/cgi-local (in order to migrate scripts from a much older server to a new one). I have edited httpd.conf to modify the following:
Code: ScriptAlias /cgi-bin/ "/var/www/cgi-local/" # # "/var/www/cgi-bin" should be changed to whatever your ScriptAliased # CGI directory exists, if you have that configured. # <Directory "/var/www/cgi-local"> [Code]...
I have also tried several variations. I left ScriptAlias at the default "/var/www/cgi-bin" and I have tried it with AllowOverride None and without the +ExecCGI after Options. I am getting a 404 error when a form has an action directed at a cgi script that is in cgi-local.
I am able to telnet to the server via a remote connection, but for some reason, it will not accept mail. ere is the bounce back email I am getting.Quote:This is the mail system at host smtp.mydomain.net.I'm sorry to have to inform you that your message could notbe delivered to one or more recipients. It's attached below.For further assistance, please send mail to postmaster.If you do so, please include this problem report. You candelete your own text from the attached returned message.The mail system
<root@mydomain.net>: temporary failure. Command output: pipe: fatal: pipe_command: execvp /usr/bin/perlbin/vendor/spamc: No such file or directory Reporting-MTA: dns; smtp.mydomain.net
I'd like to configure vsftpd server in a way to allow remote user (local) too see and edit configuration files in their ftp directory starting from dot (like .htaccess, for example). With default configuration + "local_allowed = yes" it does not appear to be possible:user can successfully upload .file but could neither see if it is in directory nor download it.
I have a folder on my drive, which I would like to try mounting as a iso9660 cd. I've tried
mount -t iso9660 131/ ~/mnt/ -o loop
but mount complains that it is a folder. Is there any way to fake it into thinking it's a dvd? The contents a friend of mine copied are in fact dvd contents from a dvd, and I'd like to be able to view without opening up the "vob" files.
The situation is say all I have is a windows machine and I remotely connected via ssh to a Linux machine. Is there a way I can mount my local CD-rom on the remote Linux machine?
I'm using ubuntu server 9.10 for a home build NAS. Everything is working great just have one more thing to figure out. I have Samba set up to access my files and I set up a recycle feature so anything deleted will get moved to a Recycled folder. (I learned this the hard way after hitting delete key by accident while browsing the shares in windows. Lost 100 GB of data)
Now it is for the most part working but the permissions on folders isn't getting set right. If I delete a file in a share I can go to Recycle bin folder and delete the file for good. But if I delete a folder I can not access that folder to delete or restore from the Recycle bin folder. I have to chmod the folder before I can do anything with it. Anything I can change to get folders deleted via windows to have the right permissions when it is moved to the Recycle bin folder?
Has anyone noticed ABRT not being able to connect to Bugzilla or the remote Core Dump analysis server? Since my upgrade during the beta phase of F15 I have been unable to file bug reports from ABRT's traces, nor has it offered me the option to file bugs to bugzilla. What it does, however, is offer me to analyse the Core Dump and backtrace of the crash using either the remote server or locally with GDB. Needless to say I'm clueless in GDB, and it requires a LOT of debug symbols found in the corresponding debug packages. But trying to use the remote analysis server always results in a message of the server being busy and to try again later. IIRC the last time I was able to file a bug through ABRT was in 13, has anyone been able to do so in 15?
I am attempting to make a shared folder for people that VPN into the network. This folder needs to be accessible to windows and mac machines. So far I have the VPN through ppptd working. I just don't know how to make a folder. I feel like this should be fairly easy. I am using Lucid Lynx server edition.
can i use one samba share with a folder showing the contents of another directory.shortcuts don't work on non ubuntu systems and it won't resolve links to files no on the share