Server :: How To Monitor That CIFS File Systems Are Mounted
Jan 14, 2010
We recently had an issue with "cat /proc/mount" telling us that a CIFS file system was mounted, even though the mount was not working correctly. So we're not sure if we can trust linux to report malfunctioning mounts, so we're planning on adding a specific file on the mounted file system, and verify the mount by reading this file from the client side (linux). If linux fails to read it, we know that the mount have failed. But before we go ahead doing this I thought I'd just hear how others are doing this sort of thing - how do you make sure that mount points are up and working?
EDIT: I just saw that I've posted in the security area, not in the server area. How do I move it?
I would like to be able to test that a network mounted cifs(samba) share is actually mounted in a script file to do backups. I want to do this so that when my automatic backups run they actually go to the remote location or fail. Currently, if there is a network problem that prevents the network share from mounting, the files simply get copied to the folder (e.g. /media/backupmount) and end up filling up my small local hard drive.
if this a simple question I apologise, I'm using a SSH connection to a remote machine which also has ubuntu installed, my remote machine is connected to a windows server, using <places> <Network> and clicking on the server, doing this mounts the server into my remote file systemWhen i look around the file system of the remote machine i'm unable to make the windows server resource available to me.I Assume it has a service file in the /dev directory, but would not know what its called or what i would have to do with it.In the mean time i've managed to connect directily to the server from my local machine, (which is probbialy a better solution) but is it possible to see the server via my remate machine?
I have a little problem with my RHEL5.5 IA64. I mounted a Windows directory with 'mount -t cifs'. After a little while someone else mounted the same windows directory in the same mountpoint. The output from 'mount -v' shows me that the same directory is mounted twice in the same mountpoints. I cannot unmount it, not even with the force option. The error is "Device or resource busy". There are no open files in the shared directory and no one using the directory or subdirectories.
I have mounted a windows network share using the gnome desktop environment, using Places -> Connect To server.The network share is OK, and I have the icon on my desktop and can see all the files.I want to be able to use this network as well in the console, so I need the mount point.What is the location on the filesystem were this networkdrive gets mounted? I find nothing in /mnt and nothing in /media also using mount to look at the registered mounts, there is no entry for the networkdrive.Nevertheless, I have this networkdrive now open in my desktop, and have an option to unmount it.I know that using the mount.cifs command you can specify the mounting point.
I have a Hitachi SimpleNET adapter (entry-level NAS device) on a Seagate FreeAgent 1TB external HDD (formatted ext3). The NAS device is connected over 100MB/s ethernet to a Netgear Wireless G router. All other devices connect using Wireless G. The NAS runs embedded Linux on an ARM processor and it runs vsftpd and Samba for file transfers.
If I transfer a large file using an FTP client the transfer maxes out at around 2.5MB/s. For my purposes that's good enough, especially considering the Wireless G bottleneck. If I transfer a file from a Windows 7 client (using samba) I get around 2.2MB/s. I know the CIFS protocol has more overhead than FTP and the difference in speed isn't that noticeable.Any combination of Ubuntu and Samba results in me getting less than 1MB/s. I've tried mounting it through Nautilus (GVFS) and /etc/fstab. FTP from this same Ubuntu client gets around 2.5MB/s.
I don't have root access on the SimpleNET to change the smb.conf. I've made a few adjustments to the mount options with no success. how to either speed up 10.04 as a Samba client or mount a folder on an FTP server locally? I've tried both curlftpfs and FUSEFTP. With curlftpfs any write operation results in an I/O error and it crashes intermittently. With FUSEFTP I never got that far and couldn't even browse the folder.
When using the following cifs mount command, mount -t smbfs -o username=username,password=password //srv/shr /usr/localfolder/and the cifs share does not exist, localfolder is mounted like d????????? ? ? ? ? ? localfolderafter a number of time , when umounting we get a kern <soft lock>Is there any way to fail the mount if the destination share does not exist, ive had a quick look through man mount but can not see a solution.
I have a line in the fstab file which automatically mounts a network drive every time I start up Ubuntu. I browse to a text file on the network drive and open it using gEdit and make changes to it. Then, when I hit the save button, a bright red warning appears:
Could not save the file [path here] gedit cannot handle file: locations in write mode. check that you typed the location correctly and try again. This also happens if I do save as. Then, after this error appears, the file actually disappears (gets deleted) from the network drive and in order to save it, I have to select save as again and type in the original filename. The line in my fstab file is:
The shares get mounted correctly and you can navigate through the directories and open files.The only problem is that it randomly starts going really slow taking 30 seconds or longer to open a directory that has 2 or 3 files in it.I have tried quite a few things to try and fix this without any luck. Its getting to the point where I am having to consider recommending that we use windows instead, which I would rather not do as I think its good for students to experience different operating systems during school.
I have 3 linux systems configured for running applications in each, named system1, system2 and system3. I have around 100 GB of space in system3 under /usr but not much being used. In System1 very less space is there but mostly hits coming here and need to have proper backup, as the system1 is quite old and not planned partitons properly. So I want to use a disk having more space for backup requirements.
I have a web application(java 5, mysql server) installed on Tomcat on Red Hat 5.Part of the functionality is to allow the upload of files to a windows share. I have mounted the share using theAny file I create here from the terminal gets sions."-rwxrwxrwx"The problem is that when the application server creates the file it gets created with the following."-rw-r--r--"How can I get files created from the server to be "-rwxrwxrwx"
I restore tape backup on my Linux server. At the time of booting the server it cannot able to read the filesystems.
I am getting the following message, Code: Your system appears to have shutdown uncleanly Forcing file system integrity check due to default setting Checking root filesystem fsch.ext3: file system has unsupported features (S) (/) e2fsck: Get a never version of e2fsck! (FAILED) *** An error occurred during the file system check. *** Dropping you to a shell; the system will reboot *** when you leave the shell. Give root password for maintenance (or type Control-D to continue):
I am still puzzled by IT guru who still install a server version of Linux using the default partition system. I am curious about what the IT guys in this forum think about this, even when the server is part of a cluster.
I need a utillity which wil monitor if disk is mounted or not. Is there a software or a script for that? I use fedora 12. Something which will send an email if disk is not mounted. I use encrypted disk, when power is off, after booting on, disk is not mounted on and I don't wan't to be automatically mounted on.
I have been running a server for 3-4 years now, and my shares have been mounting just fine. Well, the network admin looked at a backup and seen that the last date backed up was june. I got to looking around and seen that the share is not mounting. I can mount it with sudo mount -a, which tells me my syntax is correct. I get an error about IPv4 socket not opened and it is aborting the operation when I run dmesg | tail, since I can use the above command to mount later, it sounds to me like it is trying to mount before the network connection is ready.
I have done some looking over some init scripts and found that in the /etc/rc.d/init.r/netfs script it has a line that states that it is checking to see if the network is up before it starts to mount the filesystems and the such. This is set to no, my question is, can I change this option to yes and get my desired results, waiting for the network to be up before it mounts the filesystems.
I'm trying to resume copying from a mounted CIFS device to my local hdd with cURL. I tried
Code: $ curl -C - -O file://myfile and also
Code: $ curl -C <manual offset> -O file://myfile (looked up the manual offset using "$ wc -c")
This resumes copying if I cancel it eg with ^C.
But it does not work if I unmount and remount the CIFS device. cURL then ignores my given offset and continues again from start as if nothing were there without saying a word. With "-C -" the same effect.
I have one NAS device and using samba share one folder without password, how can I use command mount -t cifs to mount this share folder? I tried the below command, but always popup password checking?How should I do mount this folder without password checking?
I am trying to image about 30 laptops with WinXP, and I am using Clonezilla and DRBL for the task. We will start migration to Win7 starting Q4, so for now we are still using XP. I used a Clonezilla live USB to capture a standardized image to a CIFS/SAMBA share on the enterprise file server. The file server does not support NFS. To deploy the image, I used Virtualbox to build a VM with Centos 5.5 and then later Ubuntu 10.10. I mounted the CIFS share to /home/partimag but I found that I cannot share this CIFS mount out as NFS so I was unable to deploy the image with the image still residing on the CIFS; I had to copy the image to the VM's local drive.
Now using the DRBL live distribution, which is Debian based, I was able to obtain the image from a CIFS share and then share it out to the clients to be imaged as NFS (I think). I was able to use the DRBL live for some older computers, but since that hasn't been updated in nearly 2 years, I think it's missing some device drivers for my newer machines so it doesn't work on them -- this is why I looked at using CentOS and Ubuntu. To mount the CIFS shares, I'm using the following command:
mount -t cifs -o user@domain //share_ip_addr/share_name/folder /home/mount_point
Do I need to do something different to enable the mounted CIFS share to be shared out as a NFS share so that the clients to be imaged can see the contents from the CIFS share as a NFS share? The below image depicts my setup. The workstation has two NICs. The 10 network is the enterprise network and the 192 network is for DRBL imaging only. DRBL/Clonezilla does PXE boot and leases DHCP for the laptops. The laptops are shielded from the enterprise LAN; I am not doing any kind of NAT on on the server. The Linux VM is built with dual NICs and are set to bridged mode so they appear to be a separate NIC from the VM host on the network even though they going into the same port on the wall. [URL]
There are a number of shares on the destination system; for the purposes of this thread I used D$ and F$ (corresponding to those partitions). These shares are mounted permanently via CIFS (entries in fstab) on the source system.Today I copied an ISO image of some 3.5 GB from source (S) to destination (D). md5sum on S gave a different checksum for the source ISO than that calculated by HashCheck Shell Extension for the destination ISO. I know some would argue that I shoud use the same md5sum programm for both images.
To circumvent that I 7zipped the ISO, verified it's integrity and copied that archive from S to D. Verification of the acrchive by the Win version of 7z failed.To see if it's a protocol problem I copied both ISO and archive of ISO to another D this time using sshfs (it's an Ubuntu server). Flawless copies.Then I copied both files to another Win-based server on the same network. Flawless copies.Mystified, I checked the partition's file system integrity (NTFS) where the errors occured. Minor inconsistencies (no errors according to chkdsk). So I copied both files again, once to another partition (D:) of the original D, once to that partition causing the error in the first place (F:).
(D:): archive corrupt, checksum okay (F:): this time around both okay.
What the hell can I do to nail down the problem?! I don't even know whether it's a problem of the source system or the destination.
I'm using cifs to mount windows share.I have created one credentials file and given the path in fstab to mount at boot time. Now i want to encrypt the credentials file and place that in the fstab file.But it is not accepting.. how to use encrypted file to use in fstab,so that normal users can not watch the credentials inside the file.
I have an ntfs partition that I wish to access as a normal user(non-root). For this I did the following. As root I created a folder /windows and did a chmod 777 -R on /windows. Then I added the following line to /etc/fstab
Now, the partition is mounted alright but the problem is that when any other user (non-root) creates a files in /windows (say by executing touch newfile) the newly created file has the owner and group set as root. The non-root user can create the file and he can also delete the file, however, he cannot change the permissions of the file and also the owner:group is always set as root:root. How do I get across this problem, i.e. how do I mount a partition, so that a non-root user can also change the permissions and ownerships of the files he creates.
I've got a Seagate Blackarmor NAS which I can mount with CIFS to my Centos 5 server fine but only root can read and write to it. All other users can only read. I've tried several different mount options but results are always the same.
Specific issue: I'm trying to connect the the NAS so Bacula, a backup app, can write backups to it. "bacula" is set up as a user on the NAS. BTW, I'm pretty sure the OS on the NAS is Linux, and I can connect through windows and write fine.
I am interested in turning my home server into something that I can store backups on. I do photography and therefore have a lot of photos. I use Mac OS X for my photo editing, so it must be accessible from my Macbook. I am new when it comes network storage servers, so what would be the best solution for me to be able to backup my photos seamlessly? I would like it easy enough so others can backup files without any terminal commands and such. What would you suggest? CIFS? RAID? iSCSI?
I have 2 different mounts. One points to a local windows share(NTFS ->Samba) and the other one points to a PPTP VPN connection sharing(I belive that is NTFS too). I use "cifs" scheme in my fstab to mount these. And I use my Debian box to copy between these 2 mounts. I have started using Rsync for that purpose, I think that it works fine for now. My main problem is that it looks like Rsync cannot figure out if the files are same or not in source and target folders when I use these mounts. Most of the time Rsync copies the same files and folders over and over again even though those files and folders are on the target.
I am wondering if there is a way to make this scheme work? Being on a Vpn connection(slow) a Windows box, Rsync could have save a lot of my time if it could have recognized the files and folders that are same on both ends
Last year I was looking into fault-tolerant distributed file systems and I recall one kernel-based system that required a physical partition on each machine in the cluster, but would treat it as a single volume - ie. a write on one server would appear on the disk on all the servers.Unfortunately I didn't bookmark the specific system I was looking at, and now a year later I can't remember the details.What I don't want is NFS - a single file server with a file system mounted on various machines. What I do want is mirroring - one disk shared among multiple servers, so that if one server dies, it doesn't make any difference to the rest of them.
A bit of investigation turned up Red Hat's GFS, which kind of looks like what I want, but looks more and more like an NFS model to me. I was wondering what everyone's opinion of the various options out there were.