I have 2 servers. One is an operational CentOS + cPanel server and the other is a blank CentOS server. I want to use cPanel's backup suite to backup our customers accounts but I want to do it so the blank server is mounted to the cPanel server, rather than an FTP backup. If that makes sense. I believe from previous experience that NFS is the way to go but I'm not sure.
how i have a machine installed Red Hat Enterprise Linux Server release 5.1 with 2 modem (usb & com port) still i want to configure RAS server so some pple able to connect my server and send me some files.
This is not a regular backup. I only want to backup selective directories so personal files (photographs, documents, sourcecode) will be kept safe in case of a total system meltdown. This'll be 15GB max. Basically the digital variant of a fire resistant safe. I looked into duplicity but that requires me to install gpg keys on the target machine, which I can not do. I rather have a solution that just relies on just a working shell account and diskspace on the target server.
I thought of writing a simple script to do the following: 1. Mount remote server with sshfs 2. Mount encrypted container at remote server (LUKS, TrueCrypt?) 3. Loop over predefined directories on local machine and copy to encrypted container (rdiff-backup?)
Based on these requirements: - Target server is "dumb": only ssh access + diskspace (i.e. no installing of gpg keys) - Encrypted container should grow/shrink to fit contents - Encrypted container should be easily decryptable on any OS if you have the password - Once data leaves client server it should be encrypted: sysadmin on target server should never be able to see unencrypted data.
I already have an ubuntu backup server in my location and need this one server to be backed up remotely in another state. this other location is a helpdesk so there's a danger that they can gain access to confidential data. I'll be setting up this new server as an ftp server but need to set the ftp folder to only allow access to the backup server and me. Because its remote on the helpdesk side, they'll need some access to the file system but need to be completely blocked off from the ftp folder where all the data is at. How can I make sure I can keep them away from my data and still be able to retrieve or copy files over without permission issues between both servers?
on my server I have atMail and a intranet site being hosted using mySQL for the database locally, and obviously backing it up would be a good thing. I got the backup package from the list if you use the add/remove software called dump-0.4.etc.. can't seem to find much on how to use it, set it up, nothing. Granted I could be looking for the wrong thing which I'm sure is the case. As of now it's looking like a USB drive is going to be my only option, that will change in the future as this "project" goes further (we are a windows shop trying out Linux, and the next step is going to be adding another server that's a file server at that point we'll probably go to NAS for the backup and then transer that to a USB for offsite/secondary backup).
looking for a bit of assistance from those that know. i would like to setup an offsite backup server preferably using my current ubuntu 10.10 server.i would like something that has a client that can be used on windows PC's and that is managed so i can controll bandwidth and storage space.it will only be used for storing important files and not full backups of machines.i thought about using something like comodo backup for the clients and just setting up and FTP server but just wonderd if people know of a better solution.
I am doing a LVM replicate to another server. Example: server1.foo.com has / , /boot , swap and few LVM partitions. All are in /dev/sda disk of size 80GB. /dev/sda5 is a LVM partition which has only one vg00 and it has 2 LV's (/var and /usr) and a SAN storage connected to this server which has around 500GB of single partition(a LV partition) called /data and its under vg00 .
Now I have build another server called "server2.foo.com" with same RedHat OS version - RHEL 4 and want to import the same LVM setup at destination. Down time or unmounting the filesystem is not a problem (but am trying to reduce the down time as much as possible). Is there any way that I can take a proper snapshot of whole LVM disk and restore with same setup at destination without losing the data and the lvm file system layout?
I have windows 2003 with high memory and HDD in local network which i want to make it as backup server. there are few linux and windows clients which are all running in redhat,centos and windows 2003. As i want to setup netware backup, is there any best open source backup s/w available so that the backup would run periodically. i need to isntall backup server in my windows 2003 and backup client in other linux and windows 2003. any free backup software.? i went through zmanda but that is not freely available.
I'm looking for a free backup solution how work in client-server in both environments Linux(server) and Windows(client). in my case, i want to give a disk space quota in my Linux server for each remote windows client.
i`m trying to set up a Central Backup Server with BackupPC installed on CentOS 5.6 x86_64, My CentOS has Samba3x / Winbind integrated with Active Directory i found this nice Wiki http://wiki.centos.org/HowTos/BackupPC , to get my BackupPC installed. after installing RPMForge's repo and settin` up the priorities to the repos, http://wiki.centos.org/PackageManagement/Yum/Priorities i get the following error regarding Samba3x Conflicts , i dont wanna miss up my Samba Configuration to install BackupPC, even --skip-broken option does not work for me
I have to take two copy of incoming mails on different servers(local & remote). I am using plain server without any panel on it and using sendmail, Dovecot for mail service.Is there a way(any configurations) to take a remote backup for all incoming mails? except configuring forwarder to each mail ID.
I am playing around with the idea of being able to use a cloud or instance based service to install Ubuntu 9.10 Server. This will enable me to have remote access via SSH command line.So far, I've installed Ubuntu 9.10 Server + Ubuntu Desktop to a virtual machine. I can access this via SSH and locally via the desktop. However, in the real environment the only access I am going to have initially is via SSH.
I would like to be able to connect using Windows Remote Desktop or VNC (whichever is easier and most importantly - most secure) to the machine.. even though the desktop is on there, I need to somehow configure the remote access all from the command line.I've had a read of various forums and have been trawling support forums for days but can't find a working solution for 9.10 Server or that fits my situation above where I will not have any physical access to the desktop or machine to configure remote desktop. It all has to be done via SSH/command line.
I ended up setting up a basic linux file server as a pdc for that office. Our main office is a windows 2000 ads domain. The two offices are connected with a vpn. I only have two users at the new location so I simply have the linux and samba usernames/passwords setup manually. I would like to know if it is possible to setup a domain trust between the two sites so I don't have to create a username/password in the remote site for every user at the main office to access. I did some searching but came up empty.
I've set up a Lan-to-Lan (routed) OpenVPN tunnel. For redundancy I want to set up a second VPN tunnel on a fallback gateway/firewall on the client side. Currently, both sides (server/client) know how to route packets across each others physical LAN. So no NAT is used. When the primary gateway (fw1) is connected to the VPN server all traffic runs via the fw1 tunnel. Than when the secondary gateway (fw2) connects to the VPN server and fw1 is still connected all traffic for fw1 will be delivered to fw2 and effectively destroying traffic intended for fw1. This is of course no problem if I first shutdown (fence) fw1, than set up fw2 to use the gateway IP address from fw1 and set up the VPN tunnel to the VPN server. Effectively replacing fw1 with fw2 on the client side. However, I can't seem to find a decent howto.
I am also exploring the possibility to let both tunnels active and let OpenVPN (or another tool) decide how to route packets back and forth the different LANs. A virtual IP between two gateway's both running a VPN or something similar. This would be the preferred method of course. However, I don't know how to tackle this one but I'm pretty sure there are people out there who are happy to share their 2 cents.
How to setup remote control on Ubuntu 9 for use of windows client ?
Just for use in a home network. I've had a computer for media files for a long time and now I changed from Windows 2000 to Ubuntu 9. I the home network there is some Windows W2k and XP Home and I will use VNC or RDP for remote control of this Ubuntu desktop.
I've no experience of Linux at all and I've seen some notes about this with no luck.
I am having problems gaining access to the BackupPC admin web page. The error:
[Tue Feb 22 16:43:59 2011] [error] [client 192.168.0.2] (13)Permission denied: Could not open password file: /etc/BackupPC/htpasswd [Tue Feb 22 16:43:59 2011] [error] [client 220.127.116.11] access to /backuppc failed, reason: verification of user id 'myhtaccessuser' not configured
Why is this occuring, is there anyway of getting around this? I just cant remember originally on my old system how I set this up.
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
What's the current best practice for automounting a remote drive for automated backup? I want to use Back in Time and maintain snapshots but it can't do that remotely so I have to mount a folder outside of Back in Time. I have used sshfs from this howto in the past and it works mostly but seems to lose connection and not reconnect a lot. [URL]. Is there a more "modern" way? NFS is horribly unreliable and dog slow for me so it's OUT unless it's changed in the last year.
I have a machine on my network and that machine is a mass storage server that I will eventually use as a media server (to stream movies, videoclips and music on my home theater system). I use slackware 13 on ALL of my machines.
I am trying to automate the backup of the "/home" folder of my laptop onto the mass storage server. I currently use rsnapshot and it works great, but I would like to automate the whole process, even if I am not home or in front of my machines...
Here's what I imagined (in pseudo code):
1) Poll if server is active (up); 1.1) If not: 1.1.1) Wake up the server (WOL); 1.1.2) Wait for the server to boot; 1.1.3) Confirm the server has made it to the login prompt (normal boot); 18.104.22.168) If not, send an alarm via email;