Fedora Servers :: How To Setup GFS2 Without Cluster Suite / How To Configure Cluster.conf
Nov 6, 2010
I am trying to build GFS2 cluster with 2 or 3 Fedora 14 nodes, but I've encountered some problems from the start. First luci does not work at all in Fedora 14. There is no luci_admin and even if I manage to start luci service, I get a blank white screen when I try to open it from the browser. I've googled a bit and found that I'd might be able to setup GFS if I manage to build cluster.conf manually and start the cluster suite, but I cannot find documentation on how to create cluster.conf anywhere. If anyone knows how to setup GFS2 without a cluster suite or how to configure cluster.conf.
I am working in a project that needs to set up an Apache Web Cluster. The cluster needs to be High-availability (HA) cluster and Load-balancing cluster. Someone mentioned the use of Red Hat Cluster Suite, but, honestly, I can't figure out how it works, I haven't been able to configure it correctly. The project currently have two nodes, but we need to support at least three nodes in the cluster.
i am using the redhat cluster suite (luci and ricci) on my centos 5.3. i have 2 nodes in a cluster. when i poweroff the first node on wich a vm service is running, the service switchtes to node2. so far, so good :) but when i restart node1 the service is not failback to node1! i have created a failover domain with both nodes and priorized whre node1 has prio1 and node2 has prio2.
I have lack of understanding of CentOS in general. I have looked for a remedy on other forums and google, but haven't been able to find the answer. I have a 3 node cluster that was functioning great until I decided to go offline for awhile. My config is as follows: node 2: vh1 node 3: vh2 node 4: vh6 All nodes connect to a common shared area on an iscsi device (vguests_root)
Currently vh2 and vh6 connect great, however since putting the machines back online I can no longer connect with vh1. A dmesg command on vh1 reveals the following: GFS2: fsid=: Trying to join cluster "lock_dlm", "Cluster1:vguest_roots" GFS2: fsid=Cluster1:vguest_roots.2: Joined cluster. Now mounting FS... GFS2: fsid=Cluster1:vguest_roots.2: can't mount journal #2 GFS2: fsid=Cluster1:vguest_roots.2: there are only 2 journals (0 - 1) .....
I have configured the RHEL cluster suite in RHEL 5.3 64 bit, I have formatted a storage 100 GB LVM with GFS2 filesystem with lock_dlm protocol default journel I have used 8 I have also added this filesystem in cluster resource in GFS resource option, all work fine cluster is able to remount this GFS2 LVM on another node in time of failover. But when I was checking cluster configuration file (/etc/cluster/cluster.conf) in resource syntex it was showing fs_type=gfs so my query is if it uses fs_type=gfs then is it also mount my formatted gfs2 lvm as gfs filesystem in my cluster node rather then gfs2 filesystem. Also how do i check that mounted filesystem is gfs or gfs2, Please clear my doubt.
I am using the Redhat Cluster Suite (luci and ricci) on my centos 5.4. i have 2 nodes in a cluster.I had clustered an apache server.The service is up end running and i can stop,start and switch on all two node.The problem is when i try to simulate a fault for one node.For example:The apache resource stay on the first cluster node.If i power off the first cluster node (not halt or init 0 but take off the eletric power off), the second cluster node not take the resource.With the clustat command, the service still running on the first node.But the service is down. The first node is dead.Only one the first node is join again the cluster the resource goes up on the second node.
I was following [URL] for the cluster setup. But things dint work out. One Doubt to ask "Do I need the services cman, ccsd and rgmanager one at a time on the both machine. I have been running it through script.
I have just installed a two server cluster with ricci luci and conga on centos 5.6 32bit , both servers are vmware guests and have a shared storage disk connected to them both
with a GFS2 file system on them + fencing agents configured to work with VMware Vcenter.
(this is supported by vmware and works great on 4 other centos clusters i have been runing for 4 monthes with no CLVMD).
In this setup i used for the first time CLVMD as recommnded by RedHat so i could have the flexablitly of LVM under the GFS2 file system but , i have been getting some Strange problem with it , some times after a developer has done some IO heavy task like unziping a file or a simple TAR the load goes to 10 - 15 and no task can be killed , trying to reboot the server hangs.
After hard shutting the server every thing works ok until the next time some one does the same IO work as before.
This is my first post, and i have a question with the fence software for VMWare ESXi. The fence_vmware agent only works with ESX, and redhat (in you GIT repository) has submited a new agent called fence_vmware_ng that claims to work with ESXi. But the problem is that they do not specify the version that works with that. Anybody has test the fence_vmware_ng agent for VMWare ESXi 4.0 ?, i follow the instructions here: [URL]...and i can install the software, the API from VMWARE site, etc, but in the moment when i run the agent nothing happen, The agent connects to server, i see in logs, but the off-reboot-on operations not works. Only works status operations, that return the state of a virtual machine. I have CentOS 5.3 (fully today updated) with RHCS.
I have made a cluster between two server.In luci I can see that my cluster is green and the two nodes to.I have make an IP resource and associate it to a service : green : I can relocate the service from a node to the other one and the IP appears in the list of IP addresses The problem is that I have made the same in order to configure tomcat and postgresql and it does not work...I put my configuration only for ip and tomcat:
I am currently trying to setup a cluster using heartbeat. Unfortunatly it does not do what I want to do. I am using heartbeat 2.1.4-11 on a readhat system. I am using the the cib for configuration. What I want to do is the following: I have two Nodes. I want to run a clone resource on both nodes. It's named I want to run as a clone. This works fine.Additionally I want to have two virtual IP adresses. One should be set on Node 1 the other on Node 2. This works fine as well. I want to set dependencies between the resources. And this is where I have problems. What I want to achive: In case named is not able to run on a node, the vIP should switch to the other node (which then will have both vIPs). In case both nodes are able to run named, there should be one vIP on each node. Sounds simple, but does not work.Accourding to the heartbeat docu this an be achieved by using
In luci I can see that my cluster is green and the two nodes to. I have make an IP resource and associate it to a service : green : I can relocate the service from a node to the other one and the IP appears in the list of IP addresses
The problem is that I have made the same in order to configure tomcat and postgresql and it does not work...
I need to setup an linux cluster ..so i prefer ubuntu because of support and i personally i use ubuntu.. and can any one explain in breif ..what all the things needed to setup an ubuntu based cluster my configuration for each node will be (totally 6 nodes) core2 duo with 4 gb ram i need 4 nodes and 2 for load balancing..
I'm having a very strange problem with red hat cluster. After testing it in the lab, I tried to install a new cluster and received the following error: cman not started: Cannot start, cluster name is too long or other CCS error /usr/sbin/cman_tool: aisexec daemon didn't start I've checked the internet for that error, but nothing. I decided to to take the example from cluster.conf man, which looks like that :
And still I get the same error. I can find my servers both from DNS (FQDN and short name) and they also appear in /etc/hosts.
I'm trying to build a Fedora 12 cluster to run DHCP on top of two servers using pacemaker and drbd. The pacemaker runs great but i'm having the hardest time getting DRBD installed.
My kernel version is 126.96.36.199-99.fc12.i686.PAE First i tried to [root@dhcp-primary drbd]#yum install drbd this doesn't give me everything because apparently the kernel module for DRBD must be built from source. so i unpacked drbd-8.3.6.tar.gz and tried :
I'm not very experienced in compiling from source i'd love to make an rpm for this but after scouring the web and trying different install methods, I am still at a loss.
I've tried many, many different ways of setting up the cluster.conf file, but every time I start the cman service I get a message telling me "corosync died: could not read cluster configuration".This means nothing to me, nor can I find logs, or anything on the net about this message. I'm ultimately just trying to run a simple GFS2 config on 2 Ubuntu 9.10 desktop nodes, but I can't even get a basic cluster config going.What am I missing? I've been at this for days.
Using google with search option: cman not started: Can't find local node name in cluster.conf /usr/sbin/ cman_tool: aisexec daemon didn't start.I found this URL...I have found the config_version in cluster.conf. Unfortunately, as everyone may have noticed, english is not my native tongue so I am having trouble understanding the part "Make sure you bumped the cluster config version number". Can anyone enlightened me on what should I be doing so that I could "bump" the cluster config version?
I want to set up a cluster for scientific computing (mainly statistical stuff with R). I have a few conceptual questions. First, is there a difference between a Beowulf cluster and a cluster that has single-system image ("SSI," e.g. using openSSI or LinuxPMI)? If so, what's the difference? Second, if there is a difference between Beowulf and SSI, which one is better for scientific computing? Third, does using Eucalyptus make sense for scientific computing or is this more suitable for IO-oriented operations such as web service or databases
Can ANYONE point me in the right direction on how to use storage on multiple servers as a single cluster?I thought storage cluster was for that but, after much googling, and even more help from here, I don't think that achieves my goal. My goal is to have multiple servers share a file system, to act as somewhat of a network raid, so if node-A goes down the files are available on other nodes, and hopefully so when the capacity of the nodes are reached I can add nodes to expand the "cluster".
I need to build a 3 node web server cluster to run a php application. Since the app requires users to login (which means a session state is to be maintained), I will be sharing sessions save path, I also need to share the application directory across 3 nodes. I having trouble deciding which cluster file system to select.
I am trying to install a Red Hat cluster on my servers. In order to configure an heart bit between the servers I must give a valid multicast address. I wanted to know if there is a way to validate a multicast address before using it. For example if it is already in use i don't want to use it.
I read a little about it and there is a Perl module called IO::Socket::Multicast: [URL] There is a tool called mcasttrest.pl: [URL] Oracle use it to test multicast. But I am not sure what will do the trick in Red Hat cluster.
I am trying to do the cluster storage with the Rock Cluster storage operating system . I have install the rock cluster on the main server . and I want to connect the client with the PXE boot . when I starting the client then it boot from the PXE boot into the mode of compute node. but it ask the where is to be stored the I have give the path..then server says that not able to find the directory path.
Steps:- 1) insert-ethers 2) client is started with the PXE boot. 3) it detects the dhcp . 4) At last it demand where is to be started by cd-rom, harddisk ,nfs etc.
then I started with the nfs and I have given the lanIp of server.and server is detecting the client but client is not finding the filesystem directory. into export partition directory but it is not taking path.
PATH:- /export/rock/install it is not finding this path then it is not able to start the o/s from PXE boot..Is there any solution and manual of rock or anyother solution then u can reply to me...
I have a HA cluster, all nodes have same network supplied by one ISP. And if the error occur with ISP the system will fail. So I want to set up HA cluster with different network (supplied by different ISP) between nodes. How can I do this?
I am trying to configure redhat cluster first time but no idea how to do it. I am using rhel 5.2. I have copied cluster rpms from redhat media and tried to install them but facing alot of dependencies problem. Can some one please guide how can i do redhat clustering from start. I am doing this for configuring single node clustering for zimbra suite but it is the requirement to configure redhat cluster first. One more thing i am using two virtual machines installed by using virtual box.
I am planning to configure RedHat cluster on vmware, i dont know how to configure, i googled a lot but couldn't find satisfactory docs for configuring, can any one provide me step-by-step or good docs to go ahead.
I want to configure two node cluster for qmail-toaster?? My idea is.. if one server hardware gets failed it should transfer/migrate service to other qmail-toaster server with all settings like ... domains/users/password etc etc.