Ubuntu Servers :: Small Scale Online Backup For Ubuntu Server?
Jun 30, 2010
I just finished getting a semi-old desktop computer up and running with ubuntu server edition. It's running subversion and samba for my office of 5 people. I was wondering if there was a simple online/cloud based backup service to automatically back up my subversion repository and samba shares in case of any hardware problems?
What are the solutions to consistently backup a set of VMs stored in disk images (qcow2 or raw) on a filesystem on a lvm volume (with possible snapshots).
- Online : Ideally, I would like to send a sync to all VMs before pause and snapshot, but it seems that it's not possible with KVM (it was with XEN). The save command seemed to be an option, but it crashes some VMs. So, what do you recommend to backup these VMs online ?
- Offline : Is there a command to shutdown all vms, and start them back after snapshot/backup ?
what is the best way to scale up server set up for a high growth site? without getting a heart attack, where you can just open up a beer and watch tv as you push a red button to scale the server. lets assume the site is getting 100,000 people a day.And, in thirty days, it will be getting 9 million people a month, and in a few months it will get 9 million people a day. (I know it is a bit crazy)How would you structure a roll out?
two load balancers, two front end app servers, two database servers, and a cdn.And as traffic builds up, you would up the front end app servers, and maybe add two slave database servers? and maybe replicate the servers to east and the west to serve different parts of the world.What kind of hardware? I know it is a bit weird question, and it matters a lot on app, server hardware, server set up, style, class, lifestyle etc - I just want to see if there are any aha moments and awesome set ups that I am not aware of.
I have had trouble with my wine window. It was scaled too small and i cant reach the controls to scale it back. If i just grab the sides, it just glitches up.btw it works on different desktops, but i just want to use my own desktop because it gets a bit annoying when i have to switch user to another desktop just to use wine)
I have installed a linux server in my office to run 16 machines. Its main use will be a internal mail server but will be also running websites.
I have installed Ubuntu 9.10 server x64 and have got apache running.
I am looking for the simplest more robust solution for smtp, pop3 and imap. I have only ever used qmail before and found it a pain to configure and its getting old so I though I should probably try something new. I have not much experience with running pop3 or imap on linux so would love a suggestion on that.
Some time ago, while I was still using Ubuntu 8.10 I began accessing Internet through a proxy server. Even though I had set my system proxy settings - including authentication - to pass the proxy, my online backup service (Jungle Disk) stopped connecting. I finally figured out that I had to include proxy port in the Jungle Disk Desktop tool's configuration (like this: 192.168.0.1:8080). That worked fine until I recently upgraded to Ubuntu 10.04 - with the identical proxy setup. This time I can't get any connection at all. Here's the Jungle Disk report:
I've been using Excel from OpenOffice with moderate success for the last year. Now I've just tried to open a new spreadsheet and I'm suddenly getting a highly condensed view. The cells are about 1mm X 4mm. If I expand them by selecting and dragging, then try to copy data from another spreadsheet onto the new one, the data shrinks to the small scale -- unreadable.
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
i got an old server with a 64mb flash-disk in.this seems to be the boot-disk.would it be possible to get ubuntu server on this?there is also an ide-disk in.in my opinion it should be good to boot of the flash disk and put the /home or /srv on the ide disk.someone any ideas about this?
I have a LAMP server setup with AjaXplorer that runs locally with a static IP address. All you have to do, if you're connected to our network, is go to code...
Now I want to take the next step and make it available on the internet, so that I can access my files from anywhere in the world.
How do I go about doing this, and what precautions do I need to take?
I am going to try and replace our crappy Windows/imail server with a new ubuntu 10.04 LTS based machine. It will end up with about 400 domains on it with around 3000 users total. 96% of the domains have very light use. 4% of the domains have heavy use but do not send any bulk mail.
Can someone here recommend a mail server software package that would be easy to administrate and easy to backup/replicate? All we need is POP with webmail. We do not need IMAP or want any group/collaborative functions.
What do you usually use for anti-spam on a linux mail server? We want to keep it all open source if possible.
For hardware we are looking at an HP DL360 G5 with dual Xeons, 4G of ram, and the P400 controller with 6 SAS 15k drives. Do you think a server like this will be able to handle the mail on linux? Volume is around 400k messages a day inbound with about 10% being legit/90% being spam. The server will be dedicated to mail hosting and will not host websites.
I am trying to setup a file server for a small office but have hit a couple of hurdles, is there a step by step guide how to setup a network for windows and mac computers to use it? I had setup a share etc but once I restarted the server all the files disappeared which I had in the home folder? Also when I setup users how can I use passwords that I select as everytime I set one it encrypts it and uses that instead of my one?
I installed it on very old machine P3 1200Mhz and it works ok but I have problem with text mode because I have 75x132 aprocs. text on screen and I don't see what I type or system replay. Is there simple way to change to 25x80 or something with greater letters. I don't have GDI.Someone sad to me to do:sudo dpkg-reconfigure console-setupand it's does the job but when I restart system it's gone.
How to configure DNS server for my small network. In Fedora Linux operating system I want to use Linux as a DNS server for my small office. there are files " /etc/hosts" and "/etc/resolv.conf ".
what is the purpose of these files and in which situation we use these files? For configuring a new DNS server. how can i configure my own DNS server for my small office? Any blog or site in which they show steps to configure DNS server.
The Linux samba server would be part of the Windows domain.What's the best way to add shares for all of the students, assuming there is 200 students per folder? Is there a way to add a wildcard like $(USER) (that's logged in)//fileserver/ students/classof2011/Bill_Gates? Or would a directory path have to be created for each student?Either way is fine. I'm just curious what the proper protocol is for completing that task.
I have spent some time testing out different backup solutions for my small home office during the last weeks, but still haven't found anything that have been working out too well yet. We can definitely work with a non-GUI script if that's what it takes, if only the requirements are fulfilled: - Upload to Amazon S3 Europe. We get unbelievable slow uploading speed to US, so uploading 400+ GB of data will not be happening anytime this year... - Incremental backups - only changed files shall be uploaded or we will have a big bill from Amazon in the end of each month.. - Files should not be uploaded in one big per-folder archive. This is not efficient at all, since if we change one file in a subfolder, a huge two-digit GB sized file would have to be uploaded during next backup. Not good for economy again, or traffic overhead on our internet connection.
anyone of you could share if you have been using ubuntu for large scale website or critical mission project, say for 500.0000 secure transaction per 3 hours with 4 million users accessing server. how does ubuntu perform?
i want to run a postfix server as a backup mx, but anybody knows how can i collect the fist server mails with this one? this is multipop action but how can i do it with dovecot or any other pop collector?
I already have an ubuntu backup server in my location and need this one server to be backed up remotely in another state. this other location is a helpdesk so there's a danger that they can gain access to confidential data. I'll be setting up this new server as an ftp server but need to set the ftp folder to only allow access to the backup server and me. Because its remote on the helpdesk side, they'll need some access to the file system but need to be completely blocked off from the ftp folder where all the data is at. How can I make sure I can keep them away from my data and still be able to retrieve or copy files over without permission issues between both servers?
How is it possible to store and backup our files online? I don�t know that how this concept actually works. is there any kind of device needed while storing data at another location
I got a Web Server running great. But I need to back it up. It doesn't for some reason detect either of my USB External HD's. How can I make a TAR or something and burn it to DVD and then be able to restore using that disk.
I am currently working on managing multiple linux servers in remote locations, servers particularly user for web hosting. I need to backup data to a backup server but rsync which i currently using doesnt helps is there any tool to backup every server with out modifying it bcos there are hundreds of servers so installing a tool in every server is time consuming process.
I administer several web hosting (combined with mail relays and other services) production servers under Debian GNU/Linux. I began giving these public services two years ago via three boxes: the first is a gateway which controls traffic via iptables (it's attached to a DSL modem) between a public subnet (the DMZ) and a local network which connects several workstations. In the DMZ subnet I maintain two Pentium-III era boxes, they've grown in services since I set them up. Actually, I think I should buy new ones, but, you know, I want to save money and lenghten its life.
So, they've grown in data hosted, but I've never implemented a resilent backup system. I've set up some rsync tasks sheduled via cron jobs to copy the entire UNIX file system in each of the DMZ boxes, but I'd like to be prepared before an unexpected "real" crash of some HDD, I mean, some problem that renders a disk unusable.
AFAIK, sysadmins sync entire HD backups which are capable of recovering a system via swapping the unusable unit with the backup unit. Maybe the best fashion is to implement a RAID mirroring the unit, I'm I right? So, keeping my systems as they are, I mean, capable of using 4 parallel ATA units, what would you do? Use dump, rsync or some other way to have an operational second unit with an exact copy in a bootable second drive, in order to quickly swap it if the main unit fails?
Comes to my mind to partition a second unit (so making it bootable) and backup daily via rsync only those parts of the unix file system hierarchy which are necessary to boot a system properly. What do you think about this workaround?
I'm trying to setup a server at home, it has some practical implications, but largely it is just to take a stab at it. But I need the help of someone with more experience than I in defining exactly what I'm looking to do.
Here's what I have: old PC running Gutsy server connected to router. Several laptops at home connected via wifi to router. All laptops running either Windows or Ubuntu. Here's what I'm looking for: The server centralizes file storage for all clients. I would likely incorporate a RAID and some synchronised imaging of the files. I also want the server to create disk images of the clients hdd, regardless of client OS.There would also be some shares that would be publicly accessible (myself and friends accross the country would be able to access the same drive).
So I was thinking something like what corporate environment would be nice, you log into a profile that exists on the server. Like a dumb client...all data would be stored on the server. But I'm thinking that's more like a network boot and wouldn't work via wifi (or would it?). Also that wouldn't lend itself well to laptops used on the road in areas without net access. now I'm thinking each client would have its own locally installed OS, and they would just access networked shares. I could store sensitive files on the shares, but that wouldn't provide complete backup solution for each client.
Without rambling on anymore, anyone care to throw out some ideas? I'm really just looking to see if I can do what I want. The focus is on centrallizing files, securley backing up data and client OS's and ability to restore said images quickly.
I am writing my own anacron script that backs up my most important files from my laptop to my office pc regularly using rsync. However, sometimes my laptop is on when anacron starts the back-up job but my laptop is offline, so that it cannot rsync with my desktop. Is it possible to tell anacron to re-schedule the task to either the next time I'm online again or to some later date (preferably not the next scheduled time, but some time before)?
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
Code:
-- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version`
[code]....
It seems that MySQL can open and write to the file fine, it just can't dump
I'm trying to design an inexpensive large scale DNS server but fail to find any metrics or methods to base scalabilty.Can anyone offer information on building a stable dedicated DNS server? That might be able to scale well.
how to set up dns/bind but the issue I am hitting is that I want to just create a non internet domain setup ie t60.mysuperlan. desktop.mysuperlan. etc . Using a hosts file isn't the best as its pretty much about learning all the stuff to make it work.