Fedora Servers :: Script Can't Access Mounted Location While Logged Out
Feb 24, 2011
I want to backup some data on my Fedora box to a external Hard Disk (USB). I mounted the external HD on my box. I wrote a bash script to do that and I scheduled a cronjob to execute the script. When I am online the script executes as planned. However when I am logged out the copy does not work. I also tested this with a cifs mount (via fstab) and that does not work either. I set the script to generate some output at the end and that is OK so the script does run when I am offline. I suppose the mounted locations are not reachable while logged out, is that correct? Is there a workaround so I can reach the mounted locations while logged out?
This is the set up I have: PC downstairs by a tv, with 3TB of storage containing my media, connected to the tv too. HTPC upstairs by another tv and connected to it. A few laptops and other desktops around the house which are windows based
I want the downstairs pc to act as a file server and to run my torrent client, it is running Ubuntu desktop version and has xbmc installed too for use with the tv. The upstairs htpc has xbmc live on and will access the media from the file server. What I am looking to do is to be able to log into my ubuntu machine remotely from a laptop running windows so I can manage the files and add torrents for download etc, but for this to be a complete remote session, rather than taking control over what is already being shown on the downstairs pc, like VNC does in windows.
I have two user accounts set up on the main ubuntu machine, the admin account and a media user account which is set to go straight to xbmc after log in. Also how can I make sure that the media drives are automatically mounted to allow access if the admin user is not logged in?
After a battle with Ubuntu, Django, Apache and wsgi i could reach the website i set up from another computer via ip-adress (10.37.129.6). i then restarted the server and after booting tried to access the website from outside - permission to / denied with the usual 403 error. trying to fix that, i logged in to the server and suddenly the website was available again. typed logout on the server - no access wt. 403. logged in - website can be accessed.i somehow suspect this is some strange permission problem, but i don't have a clue where to start searching. errorlogs just contain information that a / access request has been denied.
I have server in which I have 2 HDD. One is primary and in it is Linux CentOS. Last week my server stop to work and I reboot it. But after that reboot second mounted HDD in which I backup my files aren't visible at old path. Old path is - I login via SSH and there is folder named SECOND, but now that isn't case. I want to know, how I to see in which folder to enter, which one is for second HDD? I see both, but how to enter in it? And to see what is in it (my old data)?
Device Boot Start End Blocks Id System /dev/sda1 * 1 13 104391 83 Linux /dev/sda2 14 48641 390604410 8e Linux LVM
I have an external USB had drive. It is a Buffalo drive. When it is mounted it appears in /Media/Buffalo. The problem is, every time I reboot my system it is mounted to a different location. e.g. /Media/Buffalo_1 reboot again and it is mounted to /Media/Buffalo_2
The two previous mount points always remain present as empty directories. This is creating problems I need to do away with. I need to work out how to get this to mount to the same directory every time. In opensuse 11.2 with the automount option installed I did not have this problem. It has only appeared since the install of 11.3.
I am implementing hard drive encryption. I wish to pass a key file to the crypttab from an NFS mounted location. But I could see that the disk encryption process starts very early during the booting process, before fstab is run. I could not find which script, in rc5.d, starts this service. And I am confused on how nfs mount are performed from fstab, as the network service starts at a very later stage than after fstab is called to mount the local partitions/disks. In my case, I have to wait until the nfs is mounted and then call the /dev/mapper mount (in fastab) to mount the encrypted partition.
The shares get mounted correctly and you can navigate through the directories and open files.The only problem is that it randomly starts going really slow taking 30 seconds or longer to open a directory that has 2 or 3 files in it.I have tried quite a few things to try and fix this without any luck. Its getting to the point where I am having to consider recommending that we use windows instead, which I would rather not do as I think its good for students to experience different operating systems during school.
I have the system setup and working, but I want to put it in a corner and forget about it. Problem is, nothing starts running unless I'm logged in. And if I log out, everything stops again.It's running a LAMP server and has VNC and SSH servers as well. I want all of that to start without having to log in. That way I can remote reboot without worries and connect and login with either VNC or SSH.Everything seems to have an entry in /etc/init.dIs the way it's acting normal behavior? It's a fresh install, then I installed everything I needed from the repos.
Possible Duplicate: Automatically start VNC server on startup I have a Linux server at home, and by default no-one is logged in to that box. I'd like to establish a remote desktop connection to it with VNC but this fails, unless I log in first physically. It's strange to me that VNC Viewer doesn't even ask for credentials, it just displays a "connection refused" error message. What's the best way to do this? I'm running Ubuntu 10.04 on the server.
I want to give a mounted folder /mnt/folder access so that 'root and the group test have read write access' and all other users have read access I understand most of the chmod command, the users groups world etc but where in the 'command' do you specify which 'group' or 'user' you are giving the read / write access to? in all the tutorials i've seen no where do you specify the actual group or user.
how to bind a script to a F key (F12) that will run as root even when not logged in. I have a headless server on client premises where it'd be easier for them to press F12 to run this script that will be rarely needed than to give them SSH instructions etc. I know this must be do-able, but I can't get my Google-fu on for this question. The only way that I can possibly think of doing it is to touch a file whenever that key is pressed and have the script idly checking for that file every few seconds in a loop.
I'm a total Linux noob, but I've needed a working development server for a while so I've put together an old celeron box running Ubuntu Server 10.10.The box runs fine, as does the Apache and MySQL servers, even if they did take a little while to fine tune!The problem I have is that vsftpd doesn't respond unless I'm logged on locally or via putty. As long as a local user is logged in, it's fine. If I try to connect when noone is logged in, then the connection times out waiting for the server message, and thereafter I have to login and stop / restart the vsftpd to make it work again.I'm not sure if the vsftpd is set to run on boot or on login and I have no idea how to check. Vsftpd is set to allow only local users, of which there is only one - so I can't check if it would work with any user logged in
the apche2.conf and vhost file I gave the link are the machine on LAN when site is actually hosted.When some one from internet access the site then I expect a log of IP in access.log instead of which I see the IP of machine which is working as Reverse Proxy server for all such requests.What mistake did I do above.
I'm quite new to ssh tunneling but I now want to make one of my machines at home accessible to my lan network here. I used the following command to make it available trough 127.0.0.1:5555(lo interface):
ssh -L 5555:192.168.0.15:80 me@xx.xx.xx.xx -N
now I want to make it available to eth0 on 192.168.1.40:5555 How do I do this?
When I comment out the "allow from" line, I have no access to this server at all, but when "Allow from 127.0.0.1 172.23.120" is activated, I can also access that location from other IP's (I can even access it from the internet).
What I really want is access limited to the IP's in "Allow from" because I don't want anyone accessing our subversion repo's from anywhere else.
Size Label Mount point File system 52 GB Multimedia /MM ntfs 52 GB Backup /ABackup ntfs 52 GB Extras /Extras ext4 27 GB root / ext4 60 GB home /home ext4
The problem is that I cannot access the /MM and contents. I tried Properties > Permissions and changed applied the changes to subfolders and contents too. Now I can access /MM but not the contents. All are marked with a lock logo.There are numerous folders/files.Changing the permissions individually is a hectic work.possible to do it in a command line/script?
I am trying to access a mounted secondary drive through FTP, and when I try to connect to it I am not able to see any of its contents. Any suggestions? I am using Gadmin-proftp to configure. I can point it to any other folder on the main drive and see it perfectly.
i just installed ubuntu, i am having a problem in accessing my windows files. Now as you see the 500 gb hd is mounted, but when i go to places it shows that i have 2 500 gb hds, none of them is actually the one that i have the windows system on it or my files.i looked throughout the whole web, i need your help if you know how can i access my windows files.
I have a requirement of mounting an external usb with NTFS filesystem and allowing restricting its access only to the console user(even restricting access to a single group is fine). I am able to mount the usb by using the attached fdi policy. However, I want the access to be restricted to a single user or a group. Is there anything that I'm missing here. I did search on the internet, but couldn't find anything in this regard
I have a requirement of mounting an external usb with NTFS filesystem and allowing restricting its access only to the console user(even restricting access to a single group is fine). I am able to mount the usb by using the fdi policy below. However, I want the access to be restricted to a single user or a group. Is there anything that I'm missing here. I did search on the internet, but couldn't find anything in this regard.
It's the strangest thing, I've done this on a couple othervers with no issues whatsoever... here goes:I need to mount a windows share to copy some files to it, so I used this command which gets no errors:
Code: sudo mount -t smbfs -o username=XXXXX,password=XXXXX,domain=XXXX.com //192.168.12.30/operrors /home/XXXX/scripts/operrors
I have a chrooted account setup for my ssh server. However, I am trying to allow this user read only access to access on a mounted hard-drive and more specifically a specific folder from that mounted drive. I would also like to have this drive be mounted for me in my normal environment with write access.
On a modern system--Lucid, SATA 3.0--does the location of a file on the physical disk make an appreciable difference to its access speed? If so, is there a (safe) way to put a file in a particular place on the disk?
I ask because I would like to reserve some space on disk to remain unused without messing with the partition table. My thought was to do this by using dd to create some large files (4 Gb each, or so) containing zeros.But obviously I would like to put them on the slowest part of the disk, as they won't be used for anything.
I have recently installed Debian on my NAS server. I have also configured Samba for sharing the home directory of a nas user i.e. /home/nas To this directory I have read/write from a windows machine using the nas user credentials.
When I mount my RAID partition /dev/md0p1 to the /home/nas directory, I then realize that all content in this directory (files and subfolders) is only owned by the root user. When trying to access from the windows machine the /home/nas directory, I do not have any write access, only read. I have tried both the nas and the root user credentials.
I have also attempted the change the ownership of the mounted RAID partition to the nas user with the -R recursive option, but I get for the internal files/subfolders an error "operation not supported".
How can I overcome this problem?
- Is there something not done properly in the /dev/md0 array definition (i.e. ARRAY /dev/md0 level=raid1 num-devices=2 UUID=bddf8b69:c97967b5:cb104784:7fef7cc3 )?
- Is there something not done properly in the /dev/md0p1 mounting (i.e. mount /dev/md0p1 /home/nas)?
- Should I do any extra configuration before the mounting etc?
Yesterday, I upgraded to the 10.04 verion of xubuntu. Looks fine. However, I keep my data on an NTFS partition of my dual boot laptop, and am finding that I can't access it.
I have a symbolic link to that mounted partition, and when I click on it, I can see the first level of folders, but I can't execute them (even though they seem to be set to 777) and I can't open anything...