I have a remote drive mounted on my system(ubuntu 10.04 x64), and i have the contents of that drive backed up to dropbox. the problem is, if i unmount the drive, the files disappear from dropbox. is there a way to mirror the contents of the network drive to a localfolder(preferably in such a way that all changes and file deletions are changed on the local folder instantly, but unmounting doesn't delete it all)? It looks like rsync would work, but im not sure how to make it work.
I'm new here and hope to profit from your immense linux knowledge and of course to share my own experience where I can.
I'm in a student organization and we use a file server that runs linux. I can log in through ssh and copy using scp using login and password (no rsa/dsa keys because most users are windows users using winSCP and they're lacking in computer knowledge so we don't require them to mess around with keys)
However, I don't have network access everywhere, so I'd like to make a copy on my laptop harddisk of some of the folders I use most frequently. Note that I don't need it to copy files from my pc back to the remote server so I don't need two-way sync. Deleting the local copy every time and downloading a new full copy is not an option as we are talking about several gigabytes and the download speed is limited. Normally I would use Unison, however, this requires unison to be installed on both pc's and I can't install any software on the file server so this is not an option.
Any ideas on how to achieve this? I'm reasonable knowledgable about linux so I don't mind tinkering with some config files and using command line applications.
I'm using f12. My university has recently become a mirror for fedora packages. But i'm facing a trivial problem. I have set proxy for yum so that the packages that are not available on local mirror can be downloaded from other mirrors. But then i don't know how to set no proxy for my local server. Consequently it is not using my local mirror at all. Tell me how to set no proxy for the local mirror. I want my local mirror because it's damn fast. My proxy settings are like this (they go in /etc/yum.conf):
I've just set up the local mirror for my 96 CentOS 5 workstations. The mirroring script is taken from public-mirror howto and looks like this:
#!/bin/sh rsync="/usr/bin/rsync -avHzL --delete --delay-updates" # replaced -q with -v for debuging purposes, removed --bwlimit, added -L to follow symlinks mirror=centos.politechnika.lublin.pl::CentOS # tried several different same result
I tested mirror consistency by putting the mirror addres I'm syncing with, directly into repo file. Than yumex works fine. I tried several different mirrors with the same result.
I want to setup a FAI server for which I was looking for the best method of mirroring the Debian Lenny. I want to setup a local mirror with the best method available for mirroring. If it is ftpsync, please provide me some best ways of doing it. I tried ftpsync mirroring but that was not getting properly working due to insufficient I want this mirror to be accessible in my FAI setup so that I can start the installation on multiple machines and start the updates and package installation to be done from the same local mirror.
I am having problem to collecting email from remote POP3 (all the emails for a domain is stored here) and distribute it after collecting to several users defined to Linux server. I have install postfix in Linux server for email distribution.
The issue is that my CentOS workstation is in a vlan from where the Intranet's DNS servers are unreachable. For browsing the web there is an ISA proxy server, which I presume resolves DNS for my firefox. However, wget, host, ping and aria2c fail to get any sort of DNS resolution since they're being run from command line.I have exported HTTP_PROXY value, which provides me internet access on console, but,only when I connect using IP address. It fails on name resolution.
My question is:May I redirect the DNS queries to my home PC which would be running a DNS server on a non standard port?I was thinking of putting nameserver 127.0.0.1 in /etc/resolv.conf and then put iptables rule to redirect 127.0.0.1:53 UDP to a.public.ip.address:3535 UDP..I don't know if I am shooting blanks or what, I am not very much aware of this kind of setup.My main need is to provide DNS resolution to console apps.I want to utilize my company's idle bandwidth for bulk downloads, so, using proxy, SSH tunneling through my Home PC is out of question.
As many developers probably do, I have a Windows based machine on which I run XAMPP locally to test my code and a Linux machine with Fedora as my remote server.As I sometimes use .htaccess as a way to authenticate some parts of the website, I end up having two .htaccess files: one with the local path (something line D:My_Webs) and one with my remote path (something like /var/www/html/) to the password file.I have searched high and low, but I cannot seem to find any trick so that I only have to maintain one version of the .htaccess file which can work on both Linux and on Windows machines.
I have a problem while copying files from a remote computer to my local one using the scp command. I am sure that I am using it correctly, please check it below: --- firstname.lastname@example.org:~/g4work> scp email@example.com:IndirectMethod_Spher...s/H_1.mac.root . --- What I get in return (instead of the statement saying 100% of file copied) is: --- On this machine the G4SYSTEM=Linux-g++ ---
The interesting point is that the above returned statement is one of the Environment variables set on both the machines that are necessary to work with a toolkit called Geant4. Here is what I get when I type 'printenv | grep G4' just to show you (note the statement in bold): --- G4LEVELGAMMADATA=/home/blah/geant4/geant4.9.3.p02/data/PhotonEvaporation2.0 G4INSTALL=/home/blah/geant4/geant4.9.3.p02 G4LEDATA=/home/blah/geant4/geant4.9.3.p02/data/G4EMLOW6.9 G4NEUTRONHPDATA=/home/blah/geant4/geant4.9.3.p02/data/G4NDL3.13 G4VIS_BUILD_OPENGLX_DRIVER=1 G4RADIOACTIVEDATA=/home/blah/geant4/geant4.9.3.p02/data/RadioactiveDecay3.2 G4ABLADATA=/home/blah/geant4/geant4.9.3.p02/data/G4ABLA3.0 G4LIB=/home/blah/geant4/geant4.9.3.p02/lib G4VIS_BUILD_RAYTRACERX_DRIVER=1 G4LIB_BUILD_SHARED=1 G4VIS_USE_OPENGLX=1 G4UI_USE_TCSH=1 G4VIS_USE_RAYTRACERX=1 G4REALSURFACEDATA=/home/blah/geant4/geant4.9.3.p02/data/RealSurface1.0 G4SYSTEM=Linux-g++ G4WORKDIR=/home/blah/g4work --- The other thing that I would like to mention is that these Geant4 Env. Variables are loaded each time a new (bash) shell is started as a result of the bash login script.
I would like to experiment a "green" idea of virtual desktop where multiple users are served by a single powerful machine.
I have a server running 24/7. The monitor of this machine is turned off most of the time and the OS is on the login screen.
Other users, in the same local network, use less powerful machines, which could be a thin client or an old Pentium 3 machine. They access their accounts remotely and work with the GUI as if they were sitting in front of the server. Each user sees their own desktop (different themes, screen resolution, etc.). And of course it can happen that several users could log in at the same time.
The usage is modest: mostly web browsing and the usual default applications (office, wine, gimp, etc.). In particular no games or any demanding applications. The users want to use their desktop in graphical mode only.
Question: How do we call this way of using a server? Is it possible with Ubuntu? And how to implement it?
I have a fast computer in my office and I want the person using the slow computer in the same office to boot up and see the login window (gdm) and log-in from there into the fast computer and be able to use their session on the fast computer the same time I am locally logged in to the fast computer as a different user and session.Is this best done through XDMCP? Where is a good tutorial on how to set this up?
After digging around trying to figure out how to upgrade my local apt-mirror to 10.04, I figured I would share it here so anyone else who does this, can do it with ease. It requires a few steps, and editing a couple files, but the process is relatively painless.This assumes you have an apt-mirror already running on your server. Also it assumes you're using /var/www/ubuntu as your root where all your packages are stored.
The first step is to visit http://changelogs.ubuntu.com and download the meta-release packages. Create a new directory, /var/www/ubuntu/upgrade, and drop the files there.On the client's end, a single file needs to be edited, /etc/update-manager/meta-release. Change the URI's in that file to the URI of your apt-mirror and run your update manager. Your upgrade will start, only asking if you want to change your paths in your sources.list. Agree to it, and the packages will download and configure.
I recently purchased a linux VPS from [URL] just to learn and play around with. I currently ssh into the machine via Putty from various Windows boxes. Is there anyway to throw some mp3s on the remote server and then have them play locally? Server is currently running Ubuntu Server 10.04.1.
I created a local mirror with a DVD Image, is there a way to get this mirror to 5.4 now?I already updated the updates/extras repo, but now my main-repo is outdated and my updates go wrong.my directorystructure looks like this:
/server/rpmrepo/centos/5] # ls centosplus extras os updates
I am trying to create a local mirror/repository for CentOS 5.6 for updates and network installs. I have read the page here on the site about how to do it over and over again, and I still can't figure it out. I have already created the directory, but right after that, I can't go any further. I know this is much easier than I probably think it is. Would someone mind telling me, (in beginners terms) on how to do this? I just want the 'os' package and nothing else.
I have been looking into setting up a local repository for updates etc as we have an increasing number of fedora clients/servers. Searching the web I found multiple how-to's on how to accomplish this little project. One thing that bothers me is that each how-to seems to refer to a "static URL" mirror for rsync to get the rpms.
Release upgrade ends up with error: "Error during update. A problem occurred during the update. This is usually some sort of network problem, please check your network connection and retry. The server may be overloaded. Restoring original system state". When doing do-release-upgrade -d to upgrade from karmik to lucid. I using local mirror ftp://ubuntu.snacho.ru (also have http that works but not browseable). When I change lines in /etc/apt/sources.list from local mirror to official [URL] all works fine. I don't want to download 1Gb from internet (because of traffic cost). What is wrong with local mirror ? I can communicate with its owner, but what he needs to change on the mirror ?
As the title says I'm trying to build myself a local RPM mirror. I have multiple laptops and a desktop that use Fedora 11. So I used 'rsync' to setup and sync my directories. I next went on to create my repo with the 'createrepo' function. I run my server backend as FreeBSD so I moved my data over there and setup my 'lighttpd' service.
Everything went fine until I used 'rsync' and synced up my data. Am I supposed to run 'createrepo' after each sync? If so, does anyone else use the same kinda setup, even if not FreeBSD, but a different os other than linux for their server that they run this from? I've been dealing with this for 2 weeks now and finally gave up researching and testing and thought. Not something I'm good at doing. Check my register date and my first post date.
Edit: FreeBSD doesn't have a port or unofficial port for this. I noticed it seems to be written as a python script so thought I could somehow get it to run on FreeBSD with linux emulation.
I need to write a shell script which can ready content of the folder and place files on remote FTP server. I need to make sure that a file that is already placed on remote FTP server is not attempted second time. The file names will be something like Records-2011-05-09. The files will be generated by MySQL every hour.
I have done several attempts to build a local mirror of Salckware 13.0. There are two problems: I have to wait very long for a connection, they appear error during a connection which broke it so I have to restart. Is there any other way to build mirror? I am using Alien Bob scripts. Never before happened something like that. The connections with common web pages (google, etc) work well. Maybe I should look for some other mirror than those listed in mirror-slackware-current.sh.
Running Ubuntu 9.10. In the Remote Desktop config dialog I get: "Your desktop is only reachable over the local network. Others can access your computer using the address 127.0.0.1 or tabatha.local." I understand this means only the loopback ip address is available. All my other machines show their true local ip address (e.g., 192.168.1.104) in this dialog. Thus I cannot log on to this desktop from other machines.
When I try to do a remote logon from another Ubuntu 9.10 box (or from an XP box using a VNC viewer), I get: "Connection to 192.168.1.102 has been closed." What steps are needed to make this machine show its actual ip address? All file sharing between the various machines is working properly and all windows shares back and forth between XP and 'nix, and among the the vaious XP boxes and linux boxes are available as designed.
I'm looking to convert my HTPC into a remote android dev server for my girlfriend and myself, however I want to make sure that it is possible to do what I am looking to do.Is it possible to map local devices to the VNC server (such as an android phone) so that we can work on development over VNC with phones we have on our client computers? I know its a trivial matter to map local drives over VNC, but what about non-HD devices, can I still maintain full functionality as if the device itself was plugged directly into the server?'ll be installing ubuntu again this weekend, never got around to it after my last HD failure.
I have 2 servers each one with a RAID and I want them mirror they data so if one of them goes down the other one take the job with out disruption. I've heard of multipath by I want to know it in detail or learn of more options.