Fedora Servers :: Tar Backup Of Webserver On FC10?
Jan 27, 2009
I want to backup my apache website using tar. I want to make sure all the permissions and such stay the same so that if I restore a website, it's ready to go exactly as it had been.I am using the following command and would like to know if there is anything else I should neeed to do?
tar --xattrs cvzf backup.tgz /path-to-webserver-files
Now my other question is since tar doesn't store absolute paths, when I run the restore I have to be in the root directory? In another words if I run the command in a user directory will tar do something retarded like encode the paths as ../path-to-webserver or will it just be path-to-webserver so that when I run the restore from the / folder it will automatically go into the right place?
just as the title states i need to upgrade my os from fc4 to fc10 first let me say i have no physical access to the box. all i have is root ssh access to it. So i need to know the best way of upgrading that will be safe. before i do the upgrade the server will be backed up so i will not loose any files. if the upgrade fails the data center will reload the os with what i need on it. but i would rather not have to format.
So my server running FC4 died last night and I decided to go ahead with the long-awaited upgrade to FC10 while I was rebuilding my server. I use my server for a number of things including, but not limited to: router, firewall, web server, mail server... I have a typical configuration process that I have followed since before fedora, and it has worked well for me up through FC4, but my usual config process doesn't work on FC10.
First of all, I don't want SELinux running, I didn't see an option to not install it during the FC10 setup, so how to I properly disable or uninstall it afterwards? Second, I was unable to even configure the server to route traffic from my internal network to the web, here's the process I usually go through for this:
I don't have a static IP from my ISP, so I'm not quite sure how to add the default route. I think I need to do something like "route add default gw xxx.xxx.xxx.xxx eth1" where eth1 is my external NIC, correct?
What else do I need to do to route traffic? I noticed that I wasn't even able to ping my server from the internal network even though they are on the same subnet, my server's internal NIC has an address of 192.168.7.1 and my computer on the network has an address of 192.168.7.2. If I can get this going to I have internet access again at the very least, I can move forward with the web server, email, etc.
I recently setup a file sharing server with FC10 running Samba...
I got it running perfectly and it ran great for a few weeks. I was working on the machine today and tried to reboot it and noticed that it was stuck on FC10's new boot screen. I pressed the esc key and saw that the boot process was stuck on the smb service...
I don't get any errors and the machine doesn't freeze up or anything, it just hangs on the smb service and won't continue with the boot...
Is there anyway that I can get the system to skip the smb service so that I can get into the system? I don't know what's changed or what's causing this sudden hang in the system at boot.
I tried the ubuntu live CD to attempt to go in and edit the /etc/rc5.d scripts to try and skip the smb service start but I am not sure where to look for it...
I made a big mistake on my FC10 machine-while trying to resolve a way to install vlc- I wrote the following command:"yum erase rpm lib".After restarting the machine fedora did not load... (might be funny- I laughed...).Do you know of a way top recover from this state? (I have the liveCD disk),
I have a lab of 10 computers with ips from 192.168.1.120 to 192.168.1.130 the server's Ip is 192.168.1.116When I am on client computer I type the server's Ip address on the browser and it works. All i want is that instead of entering my servers Ip I could just enter an address like: example.lan
I'm having a problem getting BIND to start after doing the FC10->FC11 upgrade. Others have been reporting some bad symbolic links, but this isn't that problem. I've searched the forums here, but I don't see anything like this anywhere.
I'm relatively new to Linux, but willing to take a plunge and climb the steep learning curve! I've recently installed Fedora 10 on a spare box I'd like to have running in the lounge by the TV as a TV recorder. Before starting out with MythTV, I would like to get the box running the basics... and hit no major snags with display / audio drivers. I've finally managed to connect to the box via VNC from my macbook over the wireless network. The Apple standard client is slow over wifi, so have started using 'Chicken of the VNC' as the client. At the server end, I intalled vnc-server via yum. I've also tweaked things in ~/.vnc/xstartup so that my default window manager over vnc is gnome.
I have a scheduled backup to run on our server at work and since the 7/12/09 it has be making 592k files instead of 10Mb files, In mysql-admin (the GUI tool) I have a stored connection for the user 'backup', the user has select and lock rights on the databases being backed up. I have a backup profile called 'backup_regular' and in the third tab along its scheduled to backup at 2 in the morning every week day. If I look at one of the small backup files generated I see the following:
-- MySQL Administrator dump 1.4 -- -- ------------------------------------------------------ -- Server version`
It seems that MySQL can open and write to the file fine, it just can't dump
Does anyone know of any decent enterprise level backup solutions for Linux? I need to backup a few servers and a bunch of desktops onto one backup server. Using rsync/tar.gz won't cut it. I need like bi-monthly full HDD backups, and things such as that, with a nice GUI interface to add/remove systems from the backup list. I need basically something similar to CommVault or Veritas. Veritas I've used before but it has its issues, such as leaving 30GB cache files. CommVault, I have no idea how much it is, and if it supports backing up to a hard drive rather than tape.
I posted this question a while ago but haven't had any luck getting it answered so I reworded everything to make it more clear. I started an Ubuntu web server on an old work laptop. The purpose of this server is to host a small file-sharing site, a forum, and a wiki. I have no prior experience with command line Ubuntu but am trying to figure this all out.
So far Ive setup the LAMP server, so apache, mysql, and php are installed. Ive also installed phpBB3 and Mediawiki. Both phpBB3 and Mediawiki are able to open from inside my LAN. I also have SSH setup and am able to use putty to manage the server from inside my LAN. Mediawiki runs slow internally but this is a secondary concern.
Externally to the LAN I can access the webserver using SSH(but sometimes it is too slow for anything that opens up such as nano or the SQL password prompt to display and the connection drops). I can also browse to the site and see the default apache welcome message. But I cannot open the mediawiki or phpbb site. Both will sit indefinitely and not open. The access log on the server shows that the request is being passed through though.
how to create an entire server from scratch using a minimal install. I know Ubuntu Server 10.04.1 comes with packages that can be included during install but for the sake of learning I only installed SSH access.
I was able to get Apache2, MySQL, PHP5 installed and running fine. What I'm trying to do next is setup DNS so I can point fragdata.com to my Ubuntu server and have other people see my test site.
I have registered the nameserver ns1.fragdata.com with my IP and updated DNS in my registrar. I followed the tutorial but "dig fragdata.com" gives me the following:
In order to pass PCI compliance I need to upgrade our webserver from php 5.3.5 to php 5.3.6. I found this: [URL] however it doesn't seem to work. My best guess is due to the fact I have 64bit server. Is there something else I can do without going through a compile?
I'm going to be launching my website soon, and I found a company to host it on one of their dedicated servers. I think I'm going to go with fedora as to OS, and well my problem is I'm having trouble finding a company to backup up my files, that both supports fedora and well is reasonably priced.
I have several machines that I'd like to backup various folders through the linux box onto DVD-RW media.I want to keep log files of what was written and when to dvd on one machine and have it automagically assign a unique serial number that I can print on the dvd in case I need to recover.I'd like a user friendly UI that I can point and click to schedule the backup and it's type.Is there a good Fedora backup application (read easy to use/understand and configure) I can use to backup machines across a network and across multiple dvd's (if needed)? The host machine is a F11 box.These are a mix of Win server 2008, win 7, win XP and several Fedora boxes.Speaking of dvd media is this a good idea and how many erase/write cycles are they good for?
I need to install a backup server in my work environment.We have a Windows 2008 server and an old DELL PowerEdge 1750 server that has no OS on it yet.I would like to install Fedora on it and then backup the Windows Server data on the Fedora server using rsync or something else to do the backups.Do you think it's a good idea ? If not what would you use to backup the Windows server data, preferably on a linux system.
Ok let's say I have Apache Webservers on 2 different machines within my network, I have http://outterABC.com setup at dyndns.org to point to my modem at home, and my router forwards Port 80 to the ServerA Machine (i.e. 192.168.0.3). I can access my webpage I setup for the Server A Machine.
But what I want to try and do is somehow access my ServerB machine's website that is on my same network. I tried something like this http://ServerB.outterABC.com and the apache page came up with something like the page wasn't available. I want to access the content of the ServerB website, but because I have only one router, i can only forward Port 80 site traffic to my ServerA machine's website. I'm sure it's a different syntax I should use but i'm just not sure what I should enter to bring up the apache root web page for the Server B website via http://outterABC.com
I tried setting up DNS A records on ServerA, but I don't think it will work with what i'm trying to do above.
I have two servers on my network One with ubuntu 9.10 server And one with openSUSE 11.2.The ubuntu server is my webserver and runs phpsysinfo and my website. On the openSuse server i have a webbased application and some files that i want people to be able to reach by using mydomain.com wich points to my ubuntu server. Is there any way to do this?
I am planning to setup my own webserver using webmin, but from what i've read i apprently need two harddrives. one for the OS & one for the data. can i just partition my harddrive into two because i dont have another hdd lying around. i also just want to note that this is my first time trying to do something like this sooo I rly want to try to set up my own server . As a side note ill list the specs of the comp i plan to use as a server. Pentium 4 2.8ghz, 1gb DDR, 80GB 7200RPM HDD.
I am trying to set up a webserver using Ubuntu server with the EHCP plugin. I have it working and can access the EHCP page by using the IP address on my local network (192.168.0.4) So far so good. I then set uo a domain using the Domain add new(custom)from the sidebar. I provided the following info:
domainname - learner Panel - me Panel password - me2 Ftp user - DACL ftp password cullompton
Great I have now got a folder structure including httpdocs for that domain, I can also use filezilla to upload webpages etc to that folder, and have uploaded a page called index.htm One problem remains - what address do I type into the web browser to see the webpage. I've tried 192.168.0.4/learner which is what I expected to work.this will only ever be used as a web server on a local network, never on the WAN.
So in my first ever attempt at Linux (and servers) I was able to set up a working web server that is visible from outside my network and created a simple PHP page thanks to this tutorial. I installed MySQL and have some experience with MS SQL but have no clue how to get started setting up databases and writing procedures in MySQL. I know all the syntax about writing scripts for MySQL (assuming it's the same as SQL) but I don't know how to get started. Do I need to be on the web server or can I do it remotely? Is there a front end GUI that everyone uses?
I upgraded webserver to new ubuntu server 10.04 (x86-64). After upgrade the increased load from 0,3 to 1,4. On webserver running phpbb, which generating slow quieres, which not before upgrade to lucid. HW conf: Intel Core i7, 8GB ram, WD Raptor 10k rpm. Week17 upgrade to new version.
I've several webserver (ubuntu 10.10 x64) on which I would like to be able to send mail like newsletter.I don't need to recieve mail on these servers, I'm using google Apps for recieving mail for the hosted domains.There's sendmail installed by default on ubuntu server..Previoulsy I was using Exim4 configured to use gmail SMTP server to send technical mails (apticron, logwatch, cron mail), but it doesn't fit for newsletter type of mailing.On my tries, I've wanted to setup Exim4 (just for the sending part, no inbox, and not using gmail stmp) which result in:
sendmail marked a uninstalled but still there and working Exim4 complaining about not being able to bind the port 25.I've removed exim4 (/etc/init.d/exim4 were not deleted...)I've reinstalled sendmail.I got this warning on installof sendmail: /etc/mail/aliases: 5 aliases, longest 21 bytes, 91 bytes total
How do I restore a working sendmail? What do I need to do so that mail sent from these servers aren't marked as spam? (not considering the content of the mail, but the server that send the mail) I see that some company used some specialized company like "grosbill.emailingoptin.com for grosbill.com" Should I specialize one domain for all newsletter ?
How to backup in such way that, data will be backed up in every 1 hour...but every time not full data;; only the data/file which have been modified within last 1 hour. Its on locally. And how to back up local data to remote system. Because I have tried with ssh-keygen also , but everytime its requiring password.
I want to backup my server that godaddy is hosting and install it on my linux system at home. In other words I want to have a clone of my server at home, like what timemachine does on mac (carries files and programs from one mac to another to make a clone).
How do I do that? I create an image file an iso file? or what? How?
I want to do this incase my server burns up over there. I don't want to reinstall all the programs and do the set ups manually again.