Software :: File Server, Backup Windows PC's Across The Network?
May 16, 2010
I have use Linux ( Suse, Ubuntu , Red Hat ) different times for different things. My newest goal is File Server. Here are the specs, I have already made the box, just choosing the OS here is what it needs to be able to do from order of most important to least.
Specs;
CPU- AMD Phenom IIx4 945
Ram- 4GB DDR 3 1600mhz
Mobo- Asus M4a79xtd EVO
Video- ATI 4650
PSU- Corsair 650w Modular
I want to setup a Linux File Server for a small windows network (around 50 users). I do know that I am gona need Smb service/pkg for that. I haven't used Samba for a while now and as per the best of my knowledge, entire communication (including usernames and passwords) between a samba server & windows client machines will be plain text. Is there any way to secure all this communication??
Secondly, if i remember correctly, MS windows wont let me mount more than one samba shares as network disk when all my shares can be accessed by different smb users with different passwords?? is there a solution to this problem? OR may be if there is any other package available for this purpose so that i wont have to use samba?
iam trying to sync file server data into backup server machine by command- rsync -avu path/of/data ipaddress-of-backup-server:/path/where/to/save after running it ask for root password and manually it is successful.but i want to make it automatic.for that i also tried cronjob and also generated authentication key but iam not successful in login automatically..anybody know how to authenticate root to login for storing data in backup server.
I am currently trying to figure out what's the best way to backup several PCs (about 5 computers each with Windows 7) from my family.
As I want the same solution for all Computer - I set up my old computer (Windows 7) and added some hard drives and there should now be enough space to backup the data of everyone from my family via the network. (Lets call this computer "Server")
But now I am wondering whats the best way to do this? What I do not want:
I do not want to start the Server each time manually when a computer tries to backup. (I thought about using WakeOnLan.. but I do not know if this is a good idea) I do not want the Server to run permanently I do not want to make the backups manually they should backup about every week automatically.
So which Software on the Computers / or the "Server" would you recommend?
Or would you eve recommend me to use Linux on the Server? If so, which Software would you use then?
I have a Linux server running Debian 5. I need to get it backed up. I have worked with tar in my past Unix days. I've read that there are some limitations to tar though since it is intended for tape.
My server would need to backup to a Windows server share. From ther it will be backed up to an offsite location. Can anyone make a recommendation for what to use? I would like to take a full backup weekly and then a incremental daily.
I have a new Ubuntu server setup to play with. I have managed to write a basic script that will backup from 1 directory to another. I would like to be able to backup the data to a windows share or nas box. how I would do this? Not too good with mounting memory sticks or anything like that.
I have a finely working CentOS server. I want to clone the complete OS (over network) so that I can use it for same functionality on several other machines..
The place I work has a web/dns server on opensuse 10 up and running hosting a few websites (our company one & a few vhost ones). The box was set up before I got here. Now we are wanting to create a new 11.2 server that is basically a backup / clone of 1st server in case it goes down.
We have a ldap server which has been running for a couple years now, and I was told to make a fallback / backup ldap server. So should the old one fail we won't be officially screwed.
is possible configure a samba server to a Backup Domain Controller in a windows 2003 Domain ? I have a Primary Domain controller Windows server 2003 , can integrate my network with a linux samba Backup Domain Controller server ?
Is there a way/command to back up all data from a Red hat Linux 4 serve[Including user rpofiles, data, group info, encrypts] either to a Red hat Linux 5.4 machine or as an Image file or manageable resource?
I have recently upgraded to Bugzilla3 and I wanted to restore my bugzilla database with my backup but when I attempt to tar -xvvzf file.tgz I get the error: gzip: stdin: not in gzip format tar: Child returned status 1 tar: Error exit delayed from previous errors
My script that creates the backup is: #!/bin/sh datestr=`date +%m-%d-%Y` bakdirpart="bugzilla.backup.$datestr" bakdir="$HOME/$bakdirpart" mkdir "$bakdir"
(cd /etc; tar cvzf $bakdir/mysql.conf.tgz mysql) (cd /etc; tar cvzf $bakdir/apache2.conf.tgz apache2) (cd /usr/share; tar cvzf $bakdir/bzreport.share.tgz bzreport) (cd /usr/share; tar cvzf $bakdir/bugzilla.share.tgz bugzilla) (cd /var/lib; tar cvzf $bakdir/mysql.hotdb.tgz mysql) (cd /var; tar cvzf $bakdir/www.tgz www) (cd "$HOME"; tar cvf "${bakdir}.tar" "$bakdirpart")
Does redhat 5.4 x_86 support bootp? I am trying to boot the storix backup bootmedia through network.They say that I can use bootp instead of DHCP.We dont have DHCP server in our environment. I could not find the bootp package. find the bootp package for redhat 5.4 x_86
This should be a quick one. I'm trying to backup a single directory and it's subdirectories on my Lucid Server to a freenas box across my network. This is what I'm using to do that..
rsync -r -a -v -z * --delete freenas:dSIBackups It almost works perfectly except for one problem. When a file is deleted at the source, this command doesn't seem to delete it on the receiving end. I assumed that the --delete would do that but aparently not. Can anyone think of a reason that this would happen?
I made a script to backup file from each host with general password in local network. This script using SSH Pass and Rsync with this
syntax: rsync --rsh="sshpass -p password ssh -l root" hostath destinationpath Everything is okay under 9.10 version until I migrate to Ubuntu 11.04, there is always give an error: rsync error: received SIGINT, SIGTERM, or SIGHUP (code 20) at rsync.c(541) [Receiver=3.0.7]
I am using bash version: GNU bash, version 4.2.8(1)-release (i686-pc-linux-gnu) and 2.6.38-8-generic kernel
I just switch back to ubuntu after running the windows for about a 6 months again, new laptop, programs needed in windows, either way I'm back. What I had setup in windows were specific files that would auto backup to my samba file server when the network was detected. I'm looking to do the same in ubuntu now. Basically I'm thinking of writing a script to backup the files, only thing I'd be stuck with is how to tell the script to run when I connect to the network at home? Is their software already designed for this.
Through the Black Friday shuffle of getting new hardware, I now have a 500TB external drive, a 1TB external drive, and an old computer I want to set up as a home server. My family has a lot of photos that are currently stored on many different computers and are not backed up, I want 500gb of space for photos, and for those photos to be backed up. That would leave the other half of the 1TB drive for assorted things like personal backups, and general file storage. I know enough how to set up Ubuntu server edition on the computer, but the options on how I can set up the storage is stumping me.
To Recap, I have 1.5TB of storage total split 1TB/500GB. I want 500GB to be used for a central storage for the 10+ computers in my house(mostly using Windows) and that 500GB would be automatically backed up. The 500GB that's left would be used for non critical files, and wouldn't be backed up.
What is the best way of backing up the files? (script once a day that copies files? Some backup program?)
Would the 500gb drive be best for backing up to(having the 1TB be where people would put the pictures) or the other way around? Does it really matter?
Any tips on the cleanest way to have this work cleanly with Windows, Linux, and Mac? How well do photo programs(Picasa, Shotwell, iPhoto) like a setup like this? Is it possible to have different programs on different machines all reference the same file system without their automatic sorting(to folders, usually by date) messing each other up?
I have an account in university on Linux machine with 10TB of free space accessible via SFTP. I would like to backup my Windows 7 x64 laptop to university. Currently I am using rsync+cygwin, but backup is pretty slow (without shadow copy) and I hate console window appearing every day on my screen when I login.
So I am looking for something like Windows Backup but with support for SFTP. Combination of tools will work too.
I am administering a live web server i want to keep a backup of the access log file without disturbing the server performance. can anyone guide me how to to this. the size of teh log file run in GB so i will need to take a daily backup
when i use rsync command to backup my image file , it shows the following error message.
bash: line 1: /usr/bin/rsync: Argument list too long rsync: connection unexpectedly closed (0 bytes received so far) [receiver] rsync error: remote command could not be run (code 126) at io.c(463) [receiver=2.6.8]
The command which i used is rsync -avrl -e ssh cms@server:/data/cms/data/images/* /mnt/Backup/Intranet_cms_backup/images
I'm looking for a free backup solution how work in client-server in both environments Linux(server) and Windows(client). in my case, i want to give a disk space quota in my Linux server for each remote windows client.
This script simply deletes files older than a certain age (in this case 7 days) from a certain location; I use it to purge old backups nightly, and it works as expected:
# delete backups older than 7 days find /mnt/backup/* -mtime +7 -exec rm -Rf {} ;
The problem is, every morning I get an email with an error message something like this:
find: `/mnt/backup/subfolder': No such file or directory
I have some issue with my amanda backup server, which is connecting with Scalar Quantum i500 via FC. I got the error as below 3 days ago. These dumps were to tape 000289. *** A TAPE ERROR OCCURRED: [No more writable valid tape found].
Normally I will load the proper tapes and run the amflush to push stuff from the holding disk to tapes manually. However this time amflush in this case did not help, Amanda immediately responded with an out of tape error again.
Meanwhile I got some errors from dmesg as well st3: Error 18 (sugg. bt 0x0, driver bt 0x0, host bt 0x0). scsi1 (0,3,0) : reservation conflict
I have a CentOS5 server with a 1tb hard drive.There is only 80gb of data on that huge drive and now I want to make a bare metal recovery backup using AcronisMy question is, how can I estimate the amount of time the backup will take and the size of the image file? Is it based on the size of my drive or is it based on the amount of data on the drive?
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
Can windows read files from a home file server with an ext4 file system? or do I have to partition the drive with the server (ext4) and an ntfs partition with the files on?
I have installed an application manager(monitoring application) on my linux server. Now, i need to have backup schedule for my application. The application itself has executive file to backup database.But when i put this file in my crontab to schedule the backup program it wont run!50 09 * * * root /opt/ME/AppManager9/bin/BackupMysqlDB.sh