Server :: Network Failover / What Is This?
Apr 23, 2010I am familiar with active to active cluster failover and active to pasive failover, but what is network failover?
View 1 RepliesI am familiar with active to active cluster failover and active to pasive failover, but what is network failover?
View 1 RepliesWhat is the current state of NFS failover (i.e. setting up two server with shared storage, with automatic failover if one fails?) I've seen a cookbook, but no details that would let me assess how well it works. There are lots of complex issues with data consistency, but the detailed information on that is years old. Our needs are fairly simple: 2 servers, a shared array, and I'm reasonably sure that we don't use locking. However we'd like failover to work reliably without loss of data.
View 3 Replies View RelatedI'm in need of some advise from you guys. I'm currently running a live production serverA, and last week it went down for a couple of hours which was really bad to say the least.
I've been thinking about building a mirror serverB that will rsync my data nightly. Now I don't want to load balance here, I just need to be able to switch to serverB when serverA goes down for any reason.
Would the best solution for this is to change my main nameserver entry when I want to switch ? I'm just curious if it will be a few hours or an instant change.
I thought I'd ask before attempting this live.
I am using Centos. I have read places that you can use Drbd + heartbeat + nfs to make a simple failover NFS server.I can't find any document that works though. I've tried 20 or so, including some Debian ones.So, does anyone have any other ideas of how to do this?Point me in the right direction please.I want 2 nodes. One to be actively serving an NFS share The other to be ready for failover. If the first one goes out, the second takes over.Meaning, the filesystem is in sync, the IP must change, and NFS must come up
View 7 Replies View RelatedWe are having a production setup where we are having one SAN storage and two RHEL machines. Now we have created a SAN LUN, say for example trylun. Now we have mounted the same SAN partition on both the machines of RHEL on the same mountpoint path say for example /trylun. After that we have installed RHEL Cluster suite to create a failover cluster.
Now we will be having one Ingres Database service for which data will be stored in SAN storage LUN mounted on both the machine say for example /trylun. When service on one machine will be down then RHEL Cluster Suite failover cluster will takeover and then it will start the same service on another node and handle the failover. Wether Ingres will run from node 1 or node 2 will not make any difference as both are using shared SAN storage i.e /trylun in our example. So same data storage will be used by both the ingres service on both the servers.
Now I have to simulate the same in my office test environment. But the problem is, in office test environment I will not have SAN server as it is additional cost. And I will have fedora operating system.
So I wanted to know is how can we create a shared file system like SAN in fedora (Is NFS a solution). And after creating the shared file system how can we create a failover cluster in fedora if we do not have Red Hat Cluster Suite.
My question is about setting up SMB and AFP failover between two servers. The plan is to have two servers both running CentOS with one acting as a primary node and one as secondary failover node. I have never set anything like this up before. In the past I have always worked with SAN's primary XSAN/StorNext. Both of which handle failover pretty much automatically. Unfortunately there isn't the budget on this job to install a SAN. Also this is only for temporary use for a week in a production office.
My thoughts where to run the two servers and use rsync on a cron tab to keep the data synchronised between the two. In an ideal world clients would log on to the primary and if that fails, seamlessly moved over to the secondary. I'm guessing however this is not possible outside of a SAN environment. So keeping the two servers synced and the clients manually moving over to the secondary manually is, I'm guessing, my only real option.
I don't have much experience in clustering. And I'm deploying a cluster system on CentOS.But I don't know how long a node failover and another node take over those resouces to continue running service is good, fast or slow? 1s, 10s or
View 2 Replies View RelatedWe have setup a High Available Cluster on two RHEL 5.4 machines with Redhat Cluster Suite (RHCS) having following configuration.
1. Both machines have Mysql server, Apache web server and Zabbix server.
2. Mysql database and web pages reside in SAN.
3. Active machine holds virtual IP and mounted shared disk.
4. We have also included a script in RHCS which takes care of starting Mysql, Apache and zabbix server on the machine which turns active when cluster switches over.
The above configuration holds good if Active machine goes down as a result of hardware failure or Reboot. What if, If any one service say Apache/Mysql/zabbix running on active hangs or become unresponsive.How can we handle this scenario ? Please advice.
I currently have 9 physical servers that I wanted to condense to 1 physical server with 7 VMs using Xen. The only issue that I have with doing this is that if the server happens to fail due to hardware problems, I am going to have a major issue. I wanted to set up a two node cluster so in the event of something happening to one of the servers, it will automatic failover to the next.
With so many ways to cluster servers, does anyone have any suggestions on the way to perform my needed task. The OS for Xen will be Red Hat and the VMs will have many versions of linux. Some of the VMs will have mysql, apache, dns, and postfix running.
I have a dual-homed Debian server running squid, but not acting as a router. Simplied network diagram is below - there are other local hops between the gateways and the Internet.
Code:
(eth0 @ 192.168.44.2) <--> (Gateway1 @ 192.168.44.1) <--> Internet
(eth1 @ 192.168.55.2) <--> (Gateway2 @ 192.168.55.1) <--> Internet
Using Gateway1 gives a very fast, but not always reliable route to the Internet. Using Gateway2 gives a slower, but more reliable route to the Internet. The server uses Gateway1 as the default gateway.
I have written a script that pings three hosts on the Internet, and if all three are down, switches the default gateway to Gateway2. This part seems to be easy, but I'd like know if there is a way of routing a ICMP/ping out eth0 to a host, with all other traffic to the host going out eth1, so I can determine if the Internet is reachable via Gateway1 again.
Have been tasked with a couple of Sunfire X2100 that I am slapping Fedora 11 onto for some high profile tasks around the office. Have two drives of the same size in each server and would like to have the two disks mirrored for redundancy. Admittedly I am new at Linux administration and am feeling over my head.
1. Can this be managed during the installation process of Fedora 11?
2. If yes, let me know the step by step please.
3. If no, I take it a cron job of rsync is going to be my best option.
4. Alternatives insights etc.
i have set up apt-cacher-ng and it is working fine.I have configured the clients using the synaptic gui, (configuration-preferences-network-set proxy...)Is there a way to configure the clients so that they use the proxy when it is available (at the office) and they fall back to using no proxy when the proxy cannot be reached?We have a lot of laptops at the office which are on the road a lot, in the office i want them to use apt-cacher-ng for updates and installation, when the users are not in the office they should be able to install software without having to change the synaptic conf...
View 2 Replies View RelatedMy friend has a server with 2 ips, 1 primary and 1 secondary/failover. He has given me a shell account and I want to use ssh to route my home http traffic through it like a socks proxy. I connect to his server using the secondary ip like this:
ssh me@secondary_ip -p port -D forwarding_port
It builds a proxy, however it uses the primary ip of the server, not the secondary ip that I logged in with. When using irssi I've bound it to the secondary ip with no problem. If I try to use the -b flag I get the error: cannot bind: Cannot assign requested address.
how I can bind the ssh tunnel to the secondary ip?
I've configuration like this :
Quote:
I try to configure my /etc/rc.local with these setting :
Quote:
But whenever i check my router after reboot, still i can't connect to inet.
We run redundant switches that two nic's on each server connect to. We also run bonding on our servers. Because we have two switches, we can't run lacp or anything. If a switch goes into a crashed state where it doesn't pass traffic but still provides link, bonding thinks the interface is still up and thus will still send traffic through it. Does anybody know a better way to configure the fail over of the interface? This would be a similar situation to somebody using a media converter.
View 1 Replies View RelatedI have a linux machine, attached to a fibre channel SAN.
We're in the testing phases and we're attempting to get all of the bugs worked out before this goes live.
If i have my host streaming data to the storage device on the san (or from the device on the san) and simulate a path failure (by shutting down one of the host's ports on the FC switch), multipath does not pick up on another path until about 45 seconds have passed.
I can verify this by watching the statistic graph (which updates once per second) on the storage system.
I see iops running along rather nicely, and then they drop to 0 for 45 seconds, then pick right up to normal again.
This is a RehHat EL 5.5 system, with qlogic HBA's.
Am i being too picky? I'd expect multipath to recover in under 30 seconds, so as to not alarm host applications running on the linux host... 45 seconds seems like a long time to wait for a disk operation to complete.
Any tips on tuning Multipath, or the qlogic card? As it is, i've got the following options in my modprobe.conf.
alias scsi_hostadapter2 qla2xxx
Code:
My multipath.conf looks like:
Code:
I've been reading up on clustering a bit and want to create simple 2 machine cluster of which the 2nd one will turn on and take over traffic at the previous machine's IP (I'm aware there will be a couple minutes of downtime while the failover server boots). I've looked into doing a simple DNS round robin, but I think both machines would need to be online all the time for this to function correctly. I've also looked into linux-HA and GlusterFS but both seem a little intense for the simple failover I'm trying to achieve. One catch if approaching with DNS round robin... the server that I want to setup with a failover also operates as a DNS server. Anyone know the simplest way to complete this? One idea I had (that seemed like a bit of manual work): Create script that runs every 60 secs through CRON on firewall (running pfsense) and have it check if master-server is up. If it's not, it will send a WOL signal to the slave, which has a mirrored image of the first. the problem I have with that is I will also need to create a script to keep both synced properly which seems almost impossible if one is shut down until failover (my thought was periodic rsyncs.) Perhaps creating a FS to direct all files to for the master/slave would be the only way around that.
View 1 Replies View RelatedI am trying to setup multipath with failover policy on openSuSE 11. I have two qla2xxx HBA's installed and they appear to be working. Here is the output of "multipath -l" command
[Code]....
While testing, I pulled one of the two connection to SAN, and the connection failed over to second HBA connection to SAN. When I plug the cable back in, it does not fall back to original connection... It stays in failed state. Also, I noticed that failed disk (sdd disk) comes back as (sdg disk), which is probably why connection does not fall back to original HBA. But, when I run "/sbin/service multipathd restart" sdg disk shows as as enabled in multipath -l...
I am familiar with windows 2008 cluster servers, and I just started testing with centos cluster. I am creating a simple 2-node cluster, for a simple ping test.
Node 1: 10.0.0.1
Node 2: 10.0.0.1
Virtual ip: 10.0.0.10
So far, I can ping a virtual ip, and manually relocate it between the nodes, but I didn't figure out, how to do this automatically. So this is my question: How can I setup the cluster, to it automatically failover the a service to another node case one node fails?
I'm new to networks and servers, been using Linux on the desktop for a while now but always relied on the company's IT guy for setting up everyting LAN-based.
Now I want to build up my home LAN, and want to do it with Linux. I've managed to set up LAMP and file share servers.
What I am looking for is information on what I need, and how to set up a server for the following tasks:Centralized Username and Password, that when the user logs into any one of the desktops in the LAN, it uses this for authentication
Something that allows this authentication to be utilized in other servers (file access, web access, router logging, etc.). Something to make it easier for continuing permissions from one service to another. e.g. I have IPCop filtering content, and it has provisions for tracking who is making which request if there is authentication going on. (optionally) to run a script for mounting Samba shares or mapped network drives so from one system to the next. For example, in whatever box somebody logs in, it mounts a server share ("smb://Myserver/users/<username>") to a local folder ("my_user_share").
So;user "fred" ="smb://Myserver/users/fred" and user "wilma" = "smb://Myserver/users/wilma" but both would find their respective one mounted under "~/my_user_share". This would be irrespective of which box they are loggin in with. If the server share location changes (new server/servername), I change it on the server so the next time they log in it points to the right place.
I guess it is similar to Window's Active Directory, though I'm not sure what it's called, how to configure it and what it is and is not capable of doing.
I have openSUSE 11.2 installed and i need to create a gateway server that allows virtual private network connections. I want to play with my friends some lan games, but we are in different networks, so i want to create this gateway server so we can connect with VPN clients to this server and play freely.
View 7 Replies View RelatedI was playing with my debian server when something went totally wrong while i was editing something on my network interface,i removed those crap that i wrote and left the network interface configuration as it was
Like for example after re-editing my network interface,it was like :
As i did a network restart, i get this error saying :
There is already a pid file /var/run/dhclient.eth0.pid with pid 2802 killed old client process, removed PID file.
What is this error and how can i fix it,because every time im re booting my server i lost my network config.
I have an opensuse 11.3 install which I want to set up as a network boot server to install Solaris 10 on a Sun Ultra 10 client. According to what I've read, this requires rarpd and tftpd which I've set up on opensuse, but also bootparamd which I can't find for 11.3. It seems it was last included with opensuse 9.2. Does anyone know if it's available, if I could use the suse 9.2 version, or any alternative?
View 3 Replies View RelatedI am trying to network a drive which is a USB drive. The directory is /media/My Book and I am at a loss on how to network it properly. From my laptop that has linux mint, I click on network and I see SFTP File Transfer on linux-8m03 but it can never seem to mount.
View 1 Replies View Relatedhow to create network policy and system policy in opensuse11.4 and domain policies also???
View 1 Replies View Relatedi have a small home network. i have laptops and workstations that my family (public) uses and an esxi box (private) that i use to test new apps for work.i need to have the public network separate from private. I have tried using two linksys routers but was unable to get the private network to access the internet. i was thinking i could use iptables with a couple of nics but I am not sure it would work. I know this could be a lot of work for someone that has never used iptables before but will give me a reason to learn it.i am sure setting up a public and private network has been done before i just don't to buy a bunch of hardware. I have a extra workstation and a bunch of nics so i would like to go that route. I am open to suggestions.
View 2 Replies View RelatedI have a Broadcom Corporation NetXtreme BCM5721 network card on my server at work.Everything works fine for a few weeks, then randomly the network connection on the server will stop working. After it stops working, I will try to reconnect with the network manager, it shows a wired connection available, it shows the "connection in progress"animation, then the "connection disabled" icon.
I uninstalled the network manager and used manual configuration, but do you think the network manager was the issue? I can't have the server disconnecting randomly every few weeks with no way to know what the real problem is. Was there an issue with the network manager with 10.04?
When I installed OpenSUSE 11.1 on my server, at some point, I gave the server a description. Whenever I connect to any of the shares on the server, this description appears. For example, on one Windows workstation, there is a share which appears in Windows Explorer as "Midwest on 'Cincinnati Corporate ISS Server (ISS0042)' (T:)".
For the life of me, I cannot find where this description, "Cincinnati Corporate ISS Server", is stored. I've search through /etc, and poked around ad nauseum in YAST, but I am missing it. Where is this description stored? What is the YAST function to change it?
I want to setup a Linux File Server for a small windows network (around 50 users). I do know that I am gona need Smb service/pkg for that. I haven't used Samba for a while now and as per the best of my knowledge, entire communication (including usernames and passwords) between a samba server & windows client machines will be plain text. Is there any way to secure all this communication??
Secondly, if i remember correctly, MS windows wont let me mount more than one samba shares as network disk when all my shares can be accessed by different smb users with different passwords?? is there a solution to this problem? OR may be if there is any other package available for this purpose so that i wont have to use samba?
I want to set up the following server in open suse:dhcpopenldapnfs (to allow users to mount their home directories from the serverI started off with the openldap server. I configured it with dc=localdomain,dc=local as its domain. As the server machine has no internet. Though when I go to add a .ldif file with the following command
Code:
ldapadd -x -D 'cn=Administrator,dc=localdomain,dc=local' -f /home/base.ldif -W
It returns this
[code]....