Ubuntu Servers :: Remote Backup Ftp Server With Confidential Data?

Feb 15, 2011

I already have an ubuntu backup server in my location and need this one server to be backed up remotely in another state. this other location is a helpdesk so there's a danger that they can gain access to confidential data. I'll be setting up this new server as an ftp server but need to set the ftp folder to only allow access to the backup server and me. Because its remote on the helpdesk side, they'll need some access to the file system but need to be completely blocked off from the ftp folder where all the data is at. How can I make sure I can keep them away from my data and still be able to retrieve or copy files over without permission issues between both servers?

View 9 Replies


Ubuntu Servers :: Best Way To Remote Backup Of Its Server Disk(mounted)?

Dec 20, 2010

Is possible to backup disk(whole) on remote server which is mounted? If yes how,

View 2 Replies View Related

General :: Clean Confidential Data From Harddisk?

Jan 21, 2010

As i said in title, we need to erase all datas from harddisk before distroy our old servers.


dd if=/dev/zero of=/dev/hda bs=1M


shred -fzuv -n 100 /dev/hda

1. Which is the best practice to erase data, dd or shred?

2. How can i execute these commands in running server, do i need to re-boot the servers as RESCUE mode using bootable cd?

View 9 Replies View Related

Ubuntu :: Rsnapshot - Automated Data Backup Of Remote File System

Aug 28, 2010

Rsnapshot is a software written in Perl to make backup of local and remote file system. The well proven rsync is behind this utility. rsnapshot does not need root user intervention to restore the data of a normal user. It does not take much space in your Backup server. It can be easily automated (scheduled) to make life easier. Just setup once and forget it configuration. Basically it takes snapshot of file system (or a part of) in regular interval such as hourly, daily, weekly and monthly.

This can be configured easily through a simple text based configuration file. The above task can be setup in a few easy steps in a few minutes. Two major tasks are configuring rsnapshot and openssh automatic login. To make the backup automatically, we need to automate the remote login in a secured way. This can be done through openssh tools. This scenario depicts backup of desktop (assuming that IP address is data to a backup server. My desktop runs on Ubuntu 10.04 and backup server runs on Debian Squeeze. [URL]

View 2 Replies View Related

Ubuntu Servers :: Dd For Remote Backup / Possible To Use It?

Oct 5, 2010

Is it possible in dd to use it for the output file to be stored at some remote location.
I do not have free space on the LVM partition whose backup I want to have via dd.

View 3 Replies View Related

Fedora Servers :: Data Backup - Will Be Backed Up In Every 1 Hour ?

Jan 8, 2009

How to backup in such way that, data will be backed up in every 1 hour...but every time not full data;; only the data/file which have been modified within last 1 hour. Its on locally. And how to back up local data to remote system. Because I have tried with ssh-keygen also , but everytime its requiring password.

View 1 Replies View Related

Fedora Servers :: Create A Backup To A Remote File System?

Jun 26, 2010

Attempting to create a backup script to copy files from one file system to a remote file system.

When I try this I get:


# tar -cf - /mnt/raid_md1 | gzip -c | ssh -i ~/.ssh/key -l user@ "cat > /mnt/backup/fileserver.md1.tar.gz"
tar: Removing leading `/' from member names
Pseudo-terminal will not be allocated because stdin is not a terminal.
ssh: Could not resolve hostname cat > /mnt/backup/fileserver.md1.tar.gz: Name or service not known


I know that the remote file system dir is RW and the access is working fine. I am stumped...

View 3 Replies View Related

Software :: Backup A Few Servers And A Bunch Of Desktops Onto One Backup Server?

May 10, 2010

Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.

View 7 Replies View Related

Ubuntu Servers :: Backup Data From The 2TB Volume To The ( Cheaper ) Other Volumes Combined?

Jan 6, 2011

All my data is stored on a separate device/partition: a 2TB USB disk with one ext4 partition. I have 3 more USB disks: 1x 1TB + 2x 500GB = 2TB right ? I want to backup my data from the 2TB volume to the (older/slower/cheaper) other volumes combined. I am thinking to use LVM to group the 3 smaller USB disks into one 2TB logical volume and use that to back up the 2TB primary volume. Do you think that is a good idea? In case of disaster I can replace the 2TB primary disk and restore from my 2TB logical volume right ?

What happens if :

I have to replace my internal hard disk and reinstall ubuntu? Can I re-attach the logical volume ?One of the physical volumes of my volume group dies ? Can I remove it and replace it with a new physical volume (bound to have other dimensions) ?Understandably I will loose my backup data?PS: running Ubuntu 10.10 .

View 5 Replies View Related

Ubuntu Servers :: Will Be Able To Recover Data 1TB Drives / Should Backup 80GB Drive As Well?

Feb 25, 2011

I am using software RAID in Ubuntu Server Edition 9.10 to mirror(RAID1) two 1TB harddrives. These are used for data storage and websites.I also have a 80GB harddrive for the operatigsystem. This drive has no backup or RAID at all. Should this drive crash and the system therefore to become no longer bootable, will I be able to recover the data the 1TB drives or should I backup the 80GB drive as well?

View 3 Replies View Related

Ubuntu Servers :: Remote Data Access Similar To WHS?

Oct 21, 2010

What app/util for Ubuntu Server 10.10 would allow me to browse and download data stored on the server via a website interface similar to Windows home server?

View 2 Replies View Related

Server :: Sync File Server Data Into Backup Server Machine By Command- Rsync -avu?

Jun 21, 2011

iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.

View 14 Replies View Related

Ubuntu Servers :: Faster Write To Disk And Backup Of Data , Putting Together A Raid5?

May 13, 2011

Here is my brief hardware and software detail in my production environment : AMD Phenom X4 3.4gHZ (Over clock to 4gHZ, 8G of Memory, 1TB 7200rpm Harddrive, Running Ubuntu server 10.10.My web production environment were pieced together 3 weeks ago.Here is my dilemma. started out with less that 40 users and now hitting 4,000 unique users per day.Now I am thinking I need faster write to disk and backup of data so I am thinking about putting together a Raid5.

I preparation for this.I have bought a new motherboard, AMD Phenom X4 3.6, and 2 more 2TB 7200rpm (Currently, I have a 2TB 7200rpm not used much)Been digging around this forum for posts with raid setup but still not sure how to seamlessly moving the some 10Gig of data from my current running prod environment once I have RAID5 installed on this new machine via the LIVE Ubuntu Server CD.

View 3 Replies View Related

CentOS 5 :: Backup Data On Server To HDD Box?

Aug 19, 2010

Now, i thinking solution: Backup Data on Server CentOS to HDD Box. When server error, i can restore Server from HDD Box.

View 1 Replies View Related

General :: Setup Remote Backup Server?

Apr 6, 2011

I have 2 servers. One is an operational CentOS + cPanel server and the other is a blank CentOS server. I want to use cPanel's backup suite to backup our customers accounts but I want to do it so the blank server is mounted to the cPanel server, rather than an FTP backup. If that makes sense. I believe from previous experience that NFS is the way to go but I'm not sure.

View 3 Replies View Related

Debian :: Encrypted Backup To Remote Dumb Server

Mar 23, 2011

This is not a regular backup. I only want to backup selective directories so personal files (photographs, documents, sourcecode) will be kept safe in case of a total system meltdown. This'll be 15GB max. Basically the digital variant of a fire resistant safe. I looked into duplicity but that requires me to install gpg keys on the target machine, which I can not do. I rather have a solution that just relies on just a working shell account and diskspace on the target server.

I thought of writing a simple script to do the following:
1. Mount remote server with sshfs
2. Mount encrypted container at remote server (LUKS, TrueCrypt?)
3. Loop over predefined directories on local machine and copy to encrypted container (rdiff-backup?)

Based on these requirements:
- Target server is "dumb": only ssh access + diskspace (i.e. no installing of gpg keys)
- Encrypted container should grow/shrink to fit contents
- Encrypted container should be easily decryptable on any OS if you have the password
- Once data leaves client server it should be encrypted: sysadmin on target server should never be able to see unencrypted data.

View 3 Replies View Related

General :: How To Backup Server For Preserving Data

Nov 10, 2010

How to backup my Linux server for preserving data in case if hard disk crashes?

View 2 Replies View Related

Server :: Unable To Copy Data To Remote Server Using Ncftp Command

Jul 16, 2011

I am unable to use ncftp command I have defined all variables used. i have to copy the data to another server FTPS. When i am executing this command it is throwing error

ncftp -u : option unknown

I am copying total script what i am executing in my server. Please some one tell me is there any pistake in using the ncftp command , or tell me some other commands to copy data to remote server

NOW=$(date +"%d-%m-%Y")


View 6 Replies View Related

Server :: MySQL Backup - Deal With Large Amounts Of Data?

Feb 15, 2011

we've been trying to become a bit more serious about backup. It seems the better way to do MySQL backup is to use the binlog. However, that binlog is huge! We seem to produce something like 10Gb per month. I'd like to copy the backup to somewhere off the server as I don't feel like there is much to be gained by just copying it to somewhere else on the server. I recently made a full backup which after compression amounted to 2.5Gb and took me 6.5 hours to copy to my own computer ... So that solution doesn't seem practical for the binlog backup.Should we rent another server somewhere? Is it possible to find a server like that really cheap? Or is there some other solution? What are other people's MySQL backup practices?

View 8 Replies View Related

Ubuntu Servers :: Access Remote Servers Via Terminal Server Client And Xserver?

May 10, 2010

i found this video, and i really want to do the same. *newbie needs to learn [URL]...my question is, what need to be installed and how?

is there any specific configuration to make it works?

and will it work if i want to connect from Ubuntu to Fedora ?

View 1 Replies View Related

Server :: Free Backup - Give A Disk Space Quota For Each Remote Windows Client

May 2, 2011

I'm looking for a free backup solution how work in client-server in both environments Linux(server) and Windows(client). in my case, i want to give a disk space quota in my Linux server for each remote windows client.

View 2 Replies View Related

General :: Backup Script Causing Data To Be Lost On Main Server- RHEL?

Feb 15, 2010

i setup the following script ro run each night (12am) as a cron on the main server:

if mount | grep -q '/home';
rsync -ranv --delete /home/;


the main server is running on a dell poweredge 2600. The script rsyncs to a virtualised duplicate running on a hp dl380. when i set this script up and begn running data started going missing from the main server. if new files had been created the files by staff they would go missing, if data was added to existing files prior to activating the script the new changes made to that file would be lost.i just cant understand why this happened. as soon as i turned the script off after a few days it was all back to normal but the data that had gone missing had gone missing.

i just wanted to know if this could be a disk read/write issue, was the script running too soon not allowing data to be written before it can be backed up? could it bee memory? i just dont know. another developed occured after a few days of all this happening, which was one of our hard disk from the main server started mis behaving and flashing amber (attention).

View 5 Replies View Related

Debian :: Used Backup-manager - Restore The Backup Data?

Feb 4, 2011

I am now preparing myself to upgrade lenny to squeeze and decided to do a backup on my system. I used backup-manager to do the job and it worked fine. how do you restore said backup data?

View 4 Replies View Related

Server :: Append A Set (or Sets) Of Data To A File(or Files) On Remote Hosts?

Dec 26, 2009

If I need to append a set (or sets) of data to a file(or files) on remote hosts what is the best mechanism by which to do that? My first thought was ssh but the command syntax to append to a remote file isn't clear to me. Can anyone point me in the right direction here?

View 4 Replies View Related

Ubuntu Servers :: Run A Postfix Server As A Backup Mx?

Sep 13, 2010

i want to run a postfix server as a backup mx, but anybody knows how can i collect the fist server mails with this one? this is multipop action but how can i do it with dovecot or any other pop collector?

View 1 Replies View Related

Ubuntu Servers :: FTP From Server To Remote File Server?

Jul 6, 2010

I support a small business which runs a headless Ubuntu Server (10.04 32bit) as a file server which is accessed by Windows machines.Although the company has it's own back-up procedure they have decided to back-up some (none sensitive) files online. The have chosen FileFactory (http://www.filefactory.com/) as the host for this. FileFactory allows files to be uploaded to their server by FTP however I do not know how to set this up on the server.

The idea, if it is possible, is to connect to FileFactory through FTP and then synchronise the data using an Rsync command.I normally access the server through Webmin and it has vsFTP installed. I can access the company's server by FTP from inside and outside of the network so I know that vsFTP is working for incomming connections however I cannot work out how to configure it to connect to the FileFactory server.

View 3 Replies View Related

Server :: How To Backup Servers And Burn To DVD?

Jan 11, 2011

I got a Web Server running great. But I need to back it up. It doesn't for some reason detect either of my USB External HD's. How can I make a TAR or something and burn it to DVD and then be able to restore using that disk.

View 6 Replies View Related

Server :: How To Backup Multiple Servers

Jun 28, 2011

I am currently working on managing multiple linux servers in remote locations, servers particularly user for web hosting. I need to backup data to a backup server but rsync which i currently using doesnt helps is there any tool to backup every server with out modifying it bcos there are hundreds of servers so installing a tool in every server is time consuming process.

View 7 Replies View Related

Server :: Production Servers HD Backup

Feb 24, 2010

I administer several web hosting (combined with mail relays and other services) production servers under Debian GNU/Linux. I began giving these public services two years ago via three boxes: the first is a gateway which controls traffic via iptables (it's attached to a DSL modem) between a public subnet (the DMZ) and a local network which connects several workstations. In the DMZ subnet I maintain two Pentium-III era boxes, they've grown in services since I set them up. Actually, I think I should buy new ones, but, you know, I want to save money and lenghten its life.

So, they've grown in data hosted, but I've never implemented a resilent backup system. I've set up some rsync tasks sheduled via cron jobs to copy the entire UNIX file system in each of the DMZ boxes, but I'd like to be prepared before an unexpected "real" crash of some HDD, I mean, some problem that renders a disk unusable.

AFAIK, sysadmins sync entire HD backups which are capable of recovering a system via swapping the unusable unit with the backup unit. Maybe the best fashion is to implement a RAID mirroring the unit, I'm I right? So, keeping my systems as they are, I mean, capable of using 4 parallel ATA units, what would you do? Use dump, rsync or some other way to have an operational second unit with an exact copy in a bootable second drive, in order to quickly swap it if the main unit fails?

Comes to my mind to partition a second unit (so making it bootable) and backup daily via rsync only those parts of the unix file system hierarchy which are necessary to boot a system properly. What do you think about this workaround?

View 6 Replies View Related

Ubuntu Servers :: Tunnel To Remote NFS Server?

Jan 11, 2011

Does anyone know the best and simplest way to do this? I'd like the share to be mounted over the tunnel on boot with as little scripting as possible and be as secure as possible without exposing more than one port to the outside. I will be trying this method: [URL]... once the tunnel is established and 'always on' NFS would take care of the file system mount obviously. Lots of the information I have been reading is not up to date it seems. Does anyone have any experience with this?

View 1 Replies View Related

Copyrights 2005-15 www.BigResource.com, All rights reserved