Server :: Cluster Storage With The Rock Cluster Storage Operating System?
Jun 8, 2010
I am trying to do the cluster storage with the Rock Cluster storage operating system . I have install the rock cluster on the main server . and I want to connect the client with the PXE boot . when I starting the client then it boot from the PXE boot into the mode of compute node. but it ask the where is to be stored the I have give the path..then server says that not able to find the directory path.
2) client is started with the PXE boot.
3) it detects the dhcp .
4) At last it demand where is to be started by cd-rom, harddisk ,nfs etc.
then I started with the nfs and I have given the lanIp of server.and server is detecting the client but client is not finding the filesystem directory. into export partition directory but it is not taking path.
/export/rock/install it is not finding this path then it is not able to start the o/s from PXE boot..Is there any solution and manual of rock or anyother solution then u can reply to me...
Have many computers at works. Not always on as students works on it from time to time. I already have a nfs server for files, but there's plenty of disk space on workstations. I wanna make use of it as a "performance booster".Which distributed file system should i use ? Looked for Plan 9 fs, but seems a bit new and "complicated" for me.
Can ANYONE point me in the right direction on how to use storage on multiple servers as a single cluster?I thought storage cluster was for that but, after much googling, and even more help from here, I don't think that achieves my goal. My goal is to have multiple servers share a file system, to act as somewhat of a network raid, so if node-A goes down the files are available on other nodes, and hopefully so when the capacity of the nodes are reached I can add nodes to expand the "cluster".
I am trying to rock cluster for the large computing. my all slave node connected with rock cluster master node. but I want to run the graphical application on the cluster node. I am not getting this point .
I am working on the beginning of implementing a two-node cluster with shared storage (GFS) and IP address. Both machines are virtual on VMware ESX 3.5, that should not make a difference, but that is the background.current status is that I have a single node cluster built with only the IP address configured within the cluster. The issue that I am having is that I have configured a service to contain only the IP address resource, however, when I go into cluster management that "service" does not register. As such, I cannot bring it online, ping it, etc. below is my cluster.conf configuration:
I am (and still) trying to create a 2 Node cluster on Centos5.2 with a Dell MD3000 as a storage. However I am getting this when I try to probe for storages in luci: An error has occured while probing storage:
I'm having a very strange problem with red hat cluster. After testing it in the lab, I tried to install a new cluster and received the following error: cman not started: Cannot start, cluster name is too long or other CCS error /usr/sbin/cman_tool: aisexec daemon didn't start I've checked the internet for that error, but nothing. I decided to to take the example from cluster.conf man, which looks like that :
And still I get the same error. I can find my servers both from DNS (FQDN and short name) and they also appear in /etc/hosts.
I'm interested in buying a new hardware for my company. The old server (now 10 years old) should be replaced with a new one. Till now, I was looking on different hardware suppliers, boards and different other places. I found a Tyan board [URL]. The hardware spec is quite interesting and the board would fullfill our claims.
how both storage devices will be supported by Ubuntu or Debian??
I am trying to build GFS2 cluster with 2 or 3 Fedora 14 nodes, but I've encountered some problems from the start. First luci does not work at all in Fedora 14. There is no luci_admin and even if I manage to start luci service, I get a blank white screen when I try to open it from the browser. I've googled a bit and found that I'd might be able to setup GFS if I manage to build cluster.conf manually and start the cluster suite, but I cannot find documentation on how to create cluster.conf anywhere. If anyone knows how to setup GFS2 without a cluster suite or how to configure cluster.conf.
I am working in a project that needs to set up an Apache Web Cluster. The cluster needs to be High-availability (HA) cluster and Load-balancing cluster. Someone mentioned the use of Red Hat Cluster Suite, but, honestly, I can't figure out how it works, I haven't been able to configure it correctly. The project currently have two nodes, but we need to support at least three nodes in the cluster.
I wanted to know how to cluster linux server. I mean any squid proxy server or apache server cluster or any other service, but I just wanted to know how to do cluster in linux. I have been searching it but there is no hope.
I tried the openssi for debian. I've tried linux from scratch but flopped trying to get mosix or kerrighed on it. Anyone got a simple way to go to get a hard drive installed single system image cluster with 32 bit machines. I'm not a new linux user but I have seemed to reach the limit of my skills. I don't like the live cds because I plan to add software later. The live cds don't fit my needs.
Does anyone have any experience with clustering KVM? I know how to do this for Xen, but I would like to know if it can also be accomplished with KVM. I am looking for a solution that is not specific for a distro.
Can a Cluster, made up of 5 Dell Poweredge computers, each with a NIC and a HDD, be a file-server that is highly available? If I go over, and break one of the computers with a sledgehammer, will the cluster continue to operate without data loss?Goes on to say how fancy the cluster is,
I'm trying to build cluster computer system using redhat linux enterprise 5.5. And I'm on the step to set up the NFS mounting. (step 3 of part VI on this paper: However, I kept getting "permission denied" when I used the command "mount -t" to mount the slave node onto the hard drive of the master node. Even when I edited /etc/fstab file to make it automatically set up during booting, I still receive the same result. By the way, I've already logged in as a root user on both slave and master nodes.
I am having an issue with LVM on a 2 node cluster. We are using powerpath for external drives. We had a request to increase /apps filesystem which is EXT3.
On the first node we did: pvcreate /dev/emcpowercx1 and /dev/emcpowercw2 Then.... vgextend apps_vg /dev/emcpowercw2 /dev/emcpowercx1 lvresize -L +60G /dev/apps_vg/apps_lv resize2fs /dev/apps_vg/apps_lv Everything went on well , we got the /apps increased. But on the second node when I do pvs.
I am getting following error: WARNING: Duplicate VG name apps_vg: RnD1W1-peb1-JWay-MyMa-WJfb-41TE-cLwvzL (created here) takes precedence over ttOYXY-dY4h-l91r-bokz-1q5c-kn3k-MCvzUX How can I proceed from here?
I have setup a MySQL High Availability Cluster with two nodes using DRBD & Heartbeat. I have successfully installed the cluster but one question which is still not solved is that Heartbeat monitors failover of network or server failure, etc.
What will happen if MySQL service gets stopped or do not start. This heartbeat-DRBD cluster does not take any action on that part. Now I want to enable this feature too which also monitors MySQL service. I have gone through MON help on web but could not find out any good step by step tutorial.
I have a HA cluster, all nodes have same network supplied by one ISP. And if the error occur with ISP the system will fail. So I want to set up HA cluster with different network (supplied by different ISP) between nodes. How can I do this?
What I did not realize was, that DLM uses the external Ethernet Interface even when talking to the local machine/node. So iptables was blocking my DLM daemon. With iptables down or the TCP port for DLM opened, cman starts, mount works.What I have here is a fibrechannel SAN which will be directly attached to several servers in the near future. Thise servers should be enabled access to a single filesystem on the SAN (shared).I heard that the right filesystem choice for this kind of setup would be GFS, because it has a Distributed Lock Manager and one FS journal for each node.
But I am having trouble setting up GFS. I have managed to create a GFS on a small testvolume (local HDD so far), but am unable to mount it. It seems that GFS/DLM needs a lot of cluster services to run, which I do not all understand / know how to correctly setup. Also: Will the lock_dlm stuff need Ethernet communications to handle file locks? And if so, will it fetch the node list from /etc/cluster/cluster.conf to determine who to talk to?
I created a cluster with two nodes and a machine for managers with luci, if a machine reboot the cluster function by transferring the resource (IP address), if forced to stop the machine (pull the plug) the cluster does not work.
I m Not able to do complete my cluster Configuration. I have done my Cluster configuration but when i m running the system-config-cluster command the cluster configuration window appear, but after closing the window it is giving me the following error [root@node1 ~]# system-config-cluster. # when the window opens, I get the below error. When I did a search everthing pointed to the 2 rpms. I verified the version of the rpms cman and openais, they are:
error code res was not 0, it was 1 error string was /usr/sbin/cman_tool: Cannot open connection to cman, is it running? When I try to issue service cman start, I get the error:
I am using Centos. I have read places that you can use Drbd + heartbeat + nfs to make a simple failover NFS server.I can't find any document that works though. I've tried 20 or so, including some Debian ones.So, does anyone have any other ideas of how to do this?Point me in the right direction please.I want 2 nodes. One to be actively serving an NFS share The other to be ready for failover. If the first one goes out, the second takes over.Meaning, the filesystem is in sync, the IP must change, and NFS must come up
what rpms should I need install for setting up redhat cluster on RHEL 5.0 I want to create two RHEL 5.0 nodes as one cluster having oracle database server installed. And please note I have created these two nodes on VMware server for testing purpose. is it possible for creating cluster of two virtual guests.
I am planning to configure RedHat cluster on vmware, i dont know how to configure, i googled a lot but couldn't find satisfactory docs for configuring, can any one provide me step-by-step or good docs to go ahead.