I have a little problem with my RHEL5.5 IA64. I mounted a Windows directory with 'mount -t cifs'. After a little while someone else mounted the same windows directory in the same mountpoint. The output from 'mount -v' shows me that the same directory is mounted twice in the same mountpoints. I cannot unmount it, not even with the force option. The error is "Device or resource busy". There are no open files in the shared directory and no one using the directory or subdirectories.
I have a DHCP/PXE server behind a firewall. It mounts partitions on the file server on the corp. network on the other side of the firewall. Every box that PXEs also mounts partitions on the main file server.
I was hoping I could change them to mount from the DHCP/PXE server, so that server could cache and cut down on the requests through the firewall, as well as the sessions that the firewall must track.But it seams a little strange to try to export directories that are simply NFS mounts on another server already.
I would like to be able to test that a network mounted cifs(samba) share is actually mounted in a script file to do backups. I want to do this so that when my automatic backups run they actually go to the remote location or fail. Currently, if there is a network problem that prevents the network share from mounting, the files simply get copied to the folder (e.g. /media/backupmount) and end up filling up my small local hard drive.
I have a Hitachi SimpleNET adapter (entry-level NAS device) on a Seagate FreeAgent 1TB external HDD (formatted ext3). The NAS device is connected over 100MB/s ethernet to a Netgear Wireless G router. All other devices connect using Wireless G. The NAS runs embedded Linux on an ARM processor and it runs vsftpd and Samba for file transfers.
If I transfer a large file using an FTP client the transfer maxes out at around 2.5MB/s. For my purposes that's good enough, especially considering the Wireless G bottleneck. If I transfer a file from a Windows 7 client (using samba) I get around 2.2MB/s. I know the CIFS protocol has more overhead than FTP and the difference in speed isn't that noticeable.Any combination of Ubuntu and Samba results in me getting less than 1MB/s. I've tried mounting it through Nautilus (GVFS) and /etc/fstab. FTP from this same Ubuntu client gets around 2.5MB/s.
I don't have root access on the SimpleNET to change the smb.conf. I've made a few adjustments to the mount options with no success. how to either speed up 10.04 as a Samba client or mount a folder on an FTP server locally? I've tried both curlftpfs and FUSEFTP. With curlftpfs any write operation results in an I/O error and it crashes intermittently. With FUSEFTP I never got that far and couldn't even browse the folder.
I have mounted a windows network share using the gnome desktop environment, using Places -> Connect To server.The network share is OK, and I have the icon on my desktop and can see all the files.I want to be able to use this network as well in the console, so I need the mount point.What is the location on the filesystem were this networkdrive gets mounted? I find nothing in /mnt and nothing in /media also using mount to look at the registered mounts, there is no entry for the networkdrive.Nevertheless, I have this networkdrive now open in my desktop, and have an option to unmount it.I know that using the mount.cifs command you can specify the mounting point.
We recently had an issue with "cat /proc/mount" telling us that a CIFS file system was mounted, even though the mount was not working correctly. So we're not sure if we can trust linux to report malfunctioning mounts, so we're planning on adding a specific file on the mounted file system, and verify the mount by reading this file from the client side (linux). If linux fails to read it, we know that the mount have failed. But before we go ahead doing this I thought I'd just hear how others are doing this sort of thing - how do you make sure that mount points are up and working?
EDIT: I just saw that I've posted in the security area, not in the server area. How do I move it?
I have a openvpn server configured and users are using from remote location. I got some errors in the /var/log/messages file as:PHP Code:
Dec 18 16:09:37system openvpn: x.x.x.x:58983 TLS Error: TLS key negotiation failed to occur within 60 seconds (check your network connectivity) Dec 18 16:09:37 system openvpn: x.x.x.x:58983 TLS Error: TLS handshake fai
I'm running 10.04 server on a Mac G5 with 2 network interfaces, one pointing to my network (192.168.0.x) and the other to a local partner network (192.9.100.x) with which we share a network volume to give/take PDF files.My client environment is MAC OS X (from 10.4 to 10.6) and until now only one client (with 2 nics) was connected to that volume: we don't have layer3 switches to do static routes over the two networks, so I decided to use my Ubuntu Server Mac to do this (it's also my new syslog server...).
Nics are configured correctly, and the local share (192.168.0.x) is well seen by everybody.But, when I mount the remote volume (192.9.100.x) to THAT shared folder, nobody is able to connect to the samba share (that now lists the remote volume directory...). The MAC OS X tells "unable to unmount the volume". IP forwarding is also activated on /etc/sysctl.confHere is a part of my smb.conf file
When using the following cifs mount command, mount -t smbfs -o username=username,password=password //srv/shr /usr/localfolder/and the cifs share does not exist, localfolder is mounted like d????????? ? ? ? ? ? localfolderafter a number of time , when umounting we get a kern <soft lock>Is there any way to fail the mount if the destination share does not exist, ive had a quick look through man mount but can not see a solution.
I have a line in the fstab file which automatically mounts a network drive every time I start up Ubuntu. I browse to a text file on the network drive and open it using gEdit and make changes to it. Then, when I hit the save button, a bright red warning appears:
Could not save the file [path here] gedit cannot handle file: locations in write mode. check that you typed the location correctly and try again. This also happens if I do save as. Then, after this error appears, the file actually disappears (gets deleted) from the network drive and in order to save it, I have to select save as again and type in the original filename. The line in my fstab file is:
The shares get mounted correctly and you can navigate through the directories and open files.The only problem is that it randomly starts going really slow taking 30 seconds or longer to open a directory that has 2 or 3 files in it.I have tried quite a few things to try and fix this without any luck. Its getting to the point where I am having to consider recommending that we use windows instead, which I would rather not do as I think its good for students to experience different operating systems during school.
if this a simple question I apologise, I'm using a SSH connection to a remote machine which also has ubuntu installed, my remote machine is connected to a windows server, using <places> <Network> and clicking on the server, doing this mounts the server into my remote file systemWhen i look around the file system of the remote machine i'm unable to make the windows server resource available to me.I Assume it has a service file in the /dev directory, but would not know what its called or what i would have to do with it.In the mean time i've managed to connect directily to the server from my local machine, (which is probbialy a better solution) but is it possible to see the server via my remate machine?
I have an external USB had drive. It is a Buffalo drive. When it is mounted it appears in /Media/Buffalo. The problem is, every time I reboot my system it is mounted to a different location. e.g. /Media/Buffalo_1 reboot again and it is mounted to /Media/Buffalo_2
The two previous mount points always remain present as empty directories. This is creating problems I need to do away with. I need to work out how to get this to mount to the same directory every time. In opensuse 11.2 with the automount option installed I did not have this problem. It has only appeared since the install of 11.3.
I am implementing hard drive encryption. I wish to pass a key file to the crypttab from an NFS mounted location. But I could see that the disk encryption process starts very early during the booting process, before fstab is run. I could not find which script, in rc5.d, starts this service. And I am confused on how nfs mount are performed from fstab, as the network service starts at a very later stage than after fstab is called to mount the local partitions/disks. In my case, I have to wait until the nfs is mounted and then call the /dev/mapper mount (in fastab) to mount the encrypted partition.
I have a "remote" server and I want to have it's X exported to my "local" laptop. Preferably, it gets done automatically from boot. I know that virtual terminals 1-6 are all text terminals and 7 is the local X. How can I get my laptop to automatically create a virtual terminal on 12 that's dedicated to receiving the X display from the server?
I know how to get it over SSH/putty. And also how to stop the local and restart the X client on 7 to use the remote. But one of the beauties of Linux is the automation and multi-user environment. It's essentially going to act as a KVM with gui replacement over IP.
I want to backup some data on my Fedora box to a external Hard Disk (USB). I mounted the external HD on my box. I wrote a bash script to do that and I scheduled a cronjob to execute the script. When I am online the script executes as planned. However when I am logged out the copy does not work. I also tested this with a cifs mount (via fstab) and that does not work either. I set the script to generate some output at the end and that is OK so the script does run when I am offline. I suppose the mounted locations are not reachable while logged out, is that correct? Is there a workaround so I can reach the mounted locations while logged out?
Am in the process of upgrading from an ancient OpenSuSE release (7.2) to 11.2. One thing I have been unable to do that worked fine under 7.2 is remotely mounting a compact flash drive from an XP machine. Worked fine for many moons on 7.2:
# mount -t cifs -o rw //xpbox/'cf (H)' /cf0 I get: mount error(12): Cannot allocate memory Other cifs mounts of hard disks work fine.
I found a posting that says this means the memory allocation error is from the XP side. It says to fiddle with the XP registry, specifically IRPStackSize. I was not confident this fix would work since there should not be anything significantly more consuming with 11.2 compared to 7.2, and indeed, I got the same error after changing the parameter to 18 and rebooting the XP machine. Any ideas? I have some suspicion that the space and parenthesis in the share name might be fouling up someone. XP forces the share name to this for some reason.
I've set up ssh passwordless logins using keygen etc.before so I know the routine.
The problem I'm currently having is setting passwordless logins when I don't have write permission to my "root" of the remote machine. More specifically the slice provided by a commercial web hosting provider. I can ssh and sftp just fine keying in the password manually but since I'm unable to create a .ssh directory in my "root" I'm unsuccessful in scripting logins. What I'm wondering is if the .ssh directory and associated security files can be placed in an alternate location such as the httpdocs directory and pass that location to ssh in a command line parameter.
I have two external hard drives mounted to my computer. They both work and are accessible. The one shows up on my desktop when it is mounted just fine. The other one does not. I went in to Configuration editor-apps-nautilus-desktop and checked the volumes_visible. That makes the other one not visible/visible but not the one I am having problems with. I read somewhere that devices must be named for it to show up and both devices are named, I even changed the name and that didn't do anything. Again, this device shows up in my computer but not on my desktop like all my other devices do. If anyone has any ideas on how I can get this external hard drive to show up.
I'm not sure if this is a bug in Squeeze beta 2 or if it's something I've overlooked. I have a Maxtor 250 GB external USB drive that I use for backups. It gets auto-mounted fine, always in the same place, and from my normal user account I can write to it, even delete directories on it if I want to. But when, from Gnome, I select the "Safely remove" option, I get an error to the effect that it can't stop the device. The weird thing is that the thing actually *is* unmounted. I've checked the mount point and it's no longer there.Is there some package I maybe should've installed but haven't? I'm not really worried about data loss, since I'm sure the drive wouldn't unmount unless it was properly synched; it's just the error message that bugs me.
How do you find the device (e.g. /dev/*) for a mounted USB drive in Linux (Ubuntu 10.04)? I'm trying to format a Cruzer USB flash drive, and when I plug it in, the icon for the mounted filesystem appears on my desktop. However, when I open GParted, it doesn't list the filesystem as an option to partition.
The recommendations I've found through Google include monitoring tail -f /var/log/messages, which they claim should list the device name when the drive is mounted, but this never happens for me. I've also read that the USB drive would usually be linked to /dev/sdb, but this appears as a broken link on my filesystem. How else would I find the device?
I've some icons of mounted devices on my desktop. I want to display only some of them which I need. All drives are mounted automatically while booting or by scripts while logging in.For example: I've 4 mounted devices and 4 icons on my desktop and I want to have 4 mounted devices but only 2 icons.There is possibility of hiding all of icons in gconf, but I want to hide only some of them.Is it possible?
Writing a bash script and I need to check if a known HD partition is mounted or not. I do not think /etc/mtab is the place to check. Would /proc/mounts always work? To make it simple like this : cat /proc/mounts | grep /dev/sdb2
[openSUSE 11.1, kernel 184.108.40.206-0.1-Default, Gnome 2.26.3] I have a problem where I have an external USB drive which mounts at boot, with /home aboard it along with several data-only partitions, and if I try to add another USB drive *after* boot, the system changes the device IDs previously given to the original drive. The already running sdb now becomes sdc. If a user or processes are attached to /home when this occurs they are left in the twilight zone and stuff starts crashing. An added weirdness is the newly added drive isn't stealing the sdb ID, the added drive becomes sdd instead, and nothing at all is left at the sdb ID.
- System boots from sda, a non-USB drive with the system as well as /swap onboard, that should be uninvolved with the problem.
- Among the 4 partitions of the boot-mounted USB drive there are 2 lux-encrypted partitions (ext3). One is /home (originally encrypted under 10.3 and added back after machine upgraded to 11.1) and another being a data directory (later encrypted under 11.1).
- The problem _may_ occur only when the additional added drive is also lux-encrypted, but this may or may not be always true as I have limited other USB stuff to test with, most of it is also lux-encrypted.