Ubuntu Servers :: Setup An Apache Web Cluster Using Red Hat Cluster Suite
Sep 1, 2011
I am working in a project that needs to set up an Apache Web Cluster. The cluster needs to be High-availability (HA) cluster and Load-balancing cluster. Someone mentioned the use of Red Hat Cluster Suite, but, honestly, I can't figure out how it works, I haven't been able to configure it correctly. The project currently have two nodes, but we need to support at least three nodes in the cluster.
I am trying to build GFS2 cluster with 2 or 3 Fedora 14 nodes, but I've encountered some problems from the start. First luci does not work at all in Fedora 14. There is no luci_admin and even if I manage to start luci service, I get a blank white screen when I try to open it from the browser. I've googled a bit and found that I'd might be able to setup GFS if I manage to build cluster.conf manually and start the cluster suite, but I cannot find documentation on how to create cluster.conf anywhere. If anyone knows how to setup GFS2 without a cluster suite or how to configure cluster.conf.
I was following [URL] for the cluster setup. But things dint work out. One Doubt to ask "Do I need the services cman, ccsd and rgmanager one at a time on the both machine. I have been running it through script.
I have just installed a two server cluster with ricci luci and conga on centos 5.6 32bit , both servers are vmware guests and have a shared storage disk connected to them both
with a GFS2 file system on them + fencing agents configured to work with VMware Vcenter.
(this is supported by vmware and works great on 4 other centos clusters i have been runing for 4 monthes with no CLVMD).
In this setup i used for the first time CLVMD as recommnded by RedHat so i could have the flexablitly of LVM under the GFS2 file system but , i have been getting some Strange problem with it , some times after a developer has done some IO heavy task like unziping a file or a simple TAR the load goes to 10 - 15 and no task can be killed , trying to reboot the server hangs.
After hard shutting the server every thing works ok until the next time some one does the same IO work as before.
I am currently trying to setup a cluster using heartbeat. Unfortunatly it does not do what I want to do. I am using heartbeat 2.1.4-11 on a readhat system. I am using the the cib for configuration. What I want to do is the following: I have two Nodes. I want to run a clone resource on both nodes. It's named I want to run as a clone. This works fine.Additionally I want to have two virtual IP adresses. One should be set on Node 1 the other on Node 2. This works fine as well. I want to set dependencies between the resources. And this is where I have problems. What I want to achive: In case named is not able to run on a node, the vIP should switch to the other node (which then will have both vIPs). In case both nodes are able to run named, there should be one vIP on each node. Sounds simple, but does not work.Accourding to the heartbeat docu this an be achieved by using
i am using the redhat cluster suite (luci and ricci) on my centos 5.3. i have 2 nodes in a cluster. when i poweroff the first node on wich a vm service is running, the service switchtes to node2. so far, so good :) but when i restart node1 the service is not failback to node1! i have created a failover domain with both nodes and priorized whre node1 has prio1 and node2 has prio2.
This is my first post, and i have a question with the fence software for VMWare ESXi. The fence_vmware agent only works with ESX, and redhat (in you GIT repository) has submited a new agent called fence_vmware_ng that claims to work with ESXi. But the problem is that they do not specify the version that works with that. Anybody has test the fence_vmware_ng agent for VMWare ESXi 4.0 ?, i follow the instructions here: [URL]...and i can install the software, the API from VMWARE site, etc, but in the moment when i run the agent nothing happen, The agent connects to server, i see in logs, but the off-reboot-on operations not works. Only works status operations, that return the state of a virtual machine. I have CentOS 5.3 (fully today updated) with RHCS.
I have made a cluster between two server.In luci I can see that my cluster is green and the two nodes to.I have make an IP resource and associate it to a service : green : I can relocate the service from a node to the other one and the IP appears in the list of IP addresses The problem is that I have made the same in order to configure tomcat and postgresql and it does not work...I put my configuration only for ip and tomcat:
I am using the Redhat Cluster Suite (luci and ricci) on my centos 5.4. i have 2 nodes in a cluster.I had clustered an apache server.The service is up end running and i can stop,start and switch on all two node.The problem is when i try to simulate a fault for one node.For example:The apache resource stay on the first cluster node.If i power off the first cluster node (not halt or init 0 but take off the eletric power off), the second cluster node not take the resource.With the clustat command, the service still running on the first node.But the service is down. The first node is dead.Only one the first node is join again the cluster the resource goes up on the second node.
I need to setup an linux cluster ..so i prefer ubuntu because of support and i personally i use ubuntu.. and can any one explain in breif ..what all the things needed to setup an ubuntu based cluster my configuration for each node will be (totally 6 nodes) core2 duo with 4 gb ram i need 4 nodes and 2 for load balancing..
In luci I can see that my cluster is green and the two nodes to. I have make an IP resource and associate it to a service : green : I can relocate the service from a node to the other one and the IP appears in the list of IP addresses
The problem is that I have made the same in order to configure tomcat and postgresql and it does not work...
I'm having a very strange problem with red hat cluster. After testing it in the lab, I tried to install a new cluster and received the following error: cman not started: Cannot start, cluster name is too long or other CCS error /usr/sbin/cman_tool: aisexec daemon didn't start I've checked the internet for that error, but nothing. I decided to to take the example from cluster.conf man, which looks like that :
And still I get the same error. I can find my servers both from DNS (FQDN and short name) and they also appear in /etc/hosts.
I am trying to do the cluster storage with the Rock Cluster storage operating system . I have install the rock cluster on the main server . and I want to connect the client with the PXE boot . when I starting the client then it boot from the PXE boot into the mode of compute node. but it ask the where is to be stored the I have give the path..then server says that not able to find the directory path.
Steps:- 1) insert-ethers 2) client is started with the PXE boot. 3) it detects the dhcp . 4) At last it demand where is to be started by cd-rom, harddisk ,nfs etc.
then I started with the nfs and I have given the lanIp of server.and server is detecting the client but client is not finding the filesystem directory. into export partition directory but it is not taking path.
PATH:- /export/rock/install it is not finding this path then it is not able to start the o/s from PXE boot..Is there any solution and manual of rock or anyother solution then u can reply to me...
I have a HA cluster, all nodes have same network supplied by one ISP. And if the error occur with ISP the system will fail. So I want to set up HA cluster with different network (supplied by different ISP) between nodes. How can I do this?
I put together a 3 PC cluster approx. 6 years ago with Redhat, at that time I was in Grad. School. Now I am in a small company and we are planning to go for a cluster setup. I have chosen to go with Centos5.
The last time around I used NFS, RSH etc. for file sharing, communication etc. Would it be OK for me to go with the same stuff or some new software/technology has come in.
The Centos cluster setup is for high-performance computing.
I am researching the CentOS Cluster Setup. Does anyone know of a guide on this or have a thread linked that I can look at? I want to build a small cluster and then be able to add to it as it grows. Mainly I want to have web hosts, data, and mail behind it. The DNS will probably be on there own.
I'm trying to build a Fedora 12 cluster to run DHCP on top of two servers using pacemaker and drbd. The pacemaker runs great but i'm having the hardest time getting DRBD installed.
My kernel version is 18.104.22.168-99.fc12.i686.PAE First i tried to [root@dhcp-primary drbd]#yum install drbd this doesn't give me everything because apparently the kernel module for DRBD must be built from source. so i unpacked drbd-8.3.6.tar.gz and tried :
I'm not very experienced in compiling from source i'd love to make an rpm for this but after scouring the web and trying different install methods, I am still at a loss.
i want to install oscar cluster on my ubuntu 8.04 , so i added its repository from http://svn.oscar.openclustergroup.or...ki/repoTesting .i added repo from Testing the development version section.when i use oscar-config --bootstrap i receive this error :
Initializing ODA: mysql ERROR: Unknown ODA type () at /usr/bin/oda line 886[code]....
I want to set up a cluster for scientific computing (mainly statistical stuff with R). I have a few conceptual questions. First, is there a difference between a Beowulf cluster and a cluster that has single-system image ("SSI," e.g. using openSSI or LinuxPMI)? If so, what's the difference? Second, if there is a difference between Beowulf and SSI, which one is better for scientific computing? Third, does using Eucalyptus make sense for scientific computing or is this more suitable for IO-oriented operations such as web service or databases
Can ANYONE point me in the right direction on how to use storage on multiple servers as a single cluster?I thought storage cluster was for that but, after much googling, and even more help from here, I don't think that achieves my goal. My goal is to have multiple servers share a file system, to act as somewhat of a network raid, so if node-A goes down the files are available on other nodes, and hopefully so when the capacity of the nodes are reached I can add nodes to expand the "cluster".
I need to build a 3 node web server cluster to run a php application. Since the app requires users to login (which means a session state is to be maintained), I will be sharing sessions save path, I also need to share the application directory across 3 nodes. I having trouble deciding which cluster file system to select.
I am trying to install a Red Hat cluster on my servers. In order to configure an heart bit between the servers I must give a valid multicast address. I wanted to know if there is a way to validate a multicast address before using it. For example if it is already in use i don't want to use it.
I read a little about it and there is a Perl module called IO::Socket::Multicast: [URL] There is a tool called mcasttrest.pl: [URL] Oracle use it to test multicast. But I am not sure what will do the trick in Red Hat cluster.
I've been reading up on clustering a bit and want to create simple 2 machine cluster of which the 2nd one will turn on and take over traffic at the previous machine's IP (I'm aware there will be a couple minutes of downtime while the failover server boots). I've looked into doing a simple DNS round robin, but I think both machines would need to be online all the time for this to function correctly. I've also looked into linux-HA and GlusterFS but both seem a little intense for the simple failover I'm trying to achieve. One catch if approaching with DNS round robin... the server that I want to setup with a failover also operates as a DNS server. Anyone know the simplest way to complete this? One idea I had (that seemed like a bit of manual work): Create script that runs every 60 secs through CRON on firewall (running pfsense) and have it check if master-server is up. If it's not, it will send a WOL signal to the slave, which has a mirrored image of the first. the problem I have with that is I will also need to create a script to keep both synced properly which seems almost impossible if one is shut down until failover (my thought was periodic rsyncs.) Perhaps creating a FS to direct all files to for the master/slave would be the only way around that.