i'm a familiar with Linux environement ( fedora 10 user ) and i got a project in a training where i have to create a cluster with two nodes where i have to set up a number of VMs that will run applications such as ( Samba, Ldap, Zimbra, ...) but i don't know how to virtualize on top of a cluster ! i would like to know how that can be done, and how is it possible to let the VMs get ressources ( RAM & CPU ) from the two nodes ??
I installed the latest version of Centos 5.4.It came with Xen version 3.0.2 I think.I am running it on a Dell R510. It supports hardware virtualization.I was able to virtualize Windows XP and Windows 2003 server on it by selecting the install from DVD option.I was having problems trying to get a PCI DekTek card seen by the virtualized Windows domains so I tried to upgrade XEN to 3.4.2.I used the repos available from Gitmo and it went easy.
The only problem is now that when I try to create a "New virtual host" it only allows me to use HTTP, FTP option and not the iso or DVD option. They are greyed out. I was wondering if anyone has had this problem and might know a work around for the Xen 3.4.2 problem. Perhaps a better location for the kernel or something.
I chose virtualization when installing Centos 5.3. The kernel I got is 2.6.18-128.el5xen My plan is to use KVM, I disable xend service. I don't need a xen enabled kernel. How do I update the kernel to a non xen one?
I have a customer who has (currently) three servers running Centos 5.5, one running Windoze 2003 and around thirty or so workstations running XP Pro.
I've just been reading about RedHat's product - RHEVD and was wondering of anyone knows if this product (or anything similar) is aavailable under CentOS.
Basically what we need to be able to do is to virtualize the XP workstations so that the staff can connect via a browser (inside the office or off-site - or for that matter anywhere) and be presented with a standard desktop.
I know I can achieve this using Citrix XenDesktop or VMWare View but I would prefer to use a FOSS solution if one exists. Having said that it has to be robust and well past Beta
I am trying to figure out how to get an almost bare metal virtualization running, and having a hard time getting it going I tryied the Virtual Machine manager, but it wont let me do full virtualization.
I have centos 5.3 with text console onl, not X server, now I try install solaris 10 at xen with full virtualization (hvm)
Now I have some like "zombie domain"
How I can to kill this? If I reboot the server it doesn't help.
No fonts found; this probably means that the fontconfig library is not correctly configured. You may need to edit the fonts.conf configuration file. More information about fontconfig can be found in the fontconfig(3) manual page and on [url] Domain installation does not appear to have been successful. If it was, you can restart your domain by running 'virsh start solhvm123'; otherwise, please restart your installation.
Do you know what this a problem?
If I chagne --vnc to --sdl and run below command on my desktop to appear windows and disappear for 1-3 seconds, virt-install -n solekhvm3 --ram 1024 --hvm --disk path=/dev/vol1/solek --sdl --cdrom /dev/cdrom --os-variant solaris10 --force
The word on all lips are "virtualization" so I decided to smell this flavor. I have a machine running CentOS 5.3 64bit and I have unlimited access to this machine, but I don't want to reach it physically. Is it possible to install and maintain a virtual medium using ssh or other remote connection? When I installed the OS I have selected the packets regarding the virtualization, but now from where can I start? Is somewhere a guide for doing this?
If you select "support virtualization" during install, you get the xen kernel installed.
Some things do not work with this kernel (e.g. nVidia drivers). So my question is whether it is possible to install both kernels (xen and ordinary) and select between them with grub at boot time?
I did try this a while ago, by first installing with xen and then manualy adding the ordinary kernel, but the ordinary kernel failed to boot (for reasons I don't now recall - sorry). Clearly, there must be differences in the "virtualization" build other than the kernel.
The alternative, if I want to play with virtualization, would be to have two entirely separate installations, but this seems like a waste of space when surely almost everything must be identical?
I can't find anything in the Centos Xen documentation about this.
Before I try again I would just like to check if anyone actually knows if this is possible, or if not why not?
well sorry I'm new in this but I just made my cluster with about 4 good machines each with 2 or more gbs of memory and dual-core processor but the question is i can install an operating system in nodes as having the frontend or not different if possible, and how i can do it because i don't now how and i don't find material and tutorials that can help me..
I have CentOS 5.5 and try to create KVM virtual machine with CentOS 5.5 as a guest. But when I exec command: virt-install --connect qemu:///system --name test --ram 1024 --file=/var/lib/libvirt/images/test.img --file-size=10 --vnc --cdrom=/var/lib/libvirt/images/CentOS-5.5-x86_64-bin-DVD-1of2.iso --vncport=10000 --accelerate --os-type=linux --nonsparse --livecd --hvm
I am trying to build GFS2 cluster with 2 or 3 Fedora 14 nodes, but I've encountered some problems from the start. First luci does not work at all in Fedora 14. There is no luci_admin and even if I manage to start luci service, I get a blank white screen when I try to open it from the browser. I've googled a bit and found that I'd might be able to setup GFS if I manage to build cluster.conf manually and start the cluster suite, but I cannot find documentation on how to create cluster.conf anywhere. If anyone knows how to setup GFS2 without a cluster suite or how to configure cluster.conf.
I'm having a very strange problem with red hat cluster. After testing it in the lab, I tried to install a new cluster and received the following error: cman not started: Cannot start, cluster name is too long or other CCS error /usr/sbin/cman_tool: aisexec daemon didn't start I've checked the internet for that error, but nothing. I decided to to take the example from cluster.conf man, which looks like that :
And still I get the same error. I can find my servers both from DNS (FQDN and short name) and they also appear in /etc/hosts.
I am working in a project that needs to set up an Apache Web Cluster. The cluster needs to be High-availability (HA) cluster and Load-balancing cluster. Someone mentioned the use of Red Hat Cluster Suite, but, honestly, I can't figure out how it works, I haven't been able to configure it correctly. The project currently have two nodes, but we need to support at least three nodes in the cluster.
What I did not realize was, that DLM uses the external Ethernet Interface even when talking to the local machine/node. So iptables was blocking my DLM daemon. With iptables down or the TCP port for DLM opened, cman starts, mount works.What I have here is a fibrechannel SAN which will be directly attached to several servers in the near future. Thise servers should be enabled access to a single filesystem on the SAN (shared).I heard that the right filesystem choice for this kind of setup would be GFS, because it has a Distributed Lock Manager and one FS journal for each node.
But I am having trouble setting up GFS. I have managed to create a GFS on a small testvolume (local HDD so far), but am unable to mount it. It seems that GFS/DLM needs a lot of cluster services to run, which I do not all understand / know how to correctly setup. Also: Will the lock_dlm stuff need Ethernet communications to handle file locks? And if so, will it fetch the node list from /etc/cluster/cluster.conf to determine who to talk to?
I created a cluster with two nodes and a machine for managers with luci, if a machine reboot the cluster function by transferring the resource (IP address), if forced to stop the machine (pull the plug) the cluster does not work.
I have lack of understanding of CentOS in general. I have looked for a remedy on other forums and google, but haven't been able to find the answer. I have a 3 node cluster that was functioning great until I decided to go offline for awhile. My config is as follows: node 2: vh1 node 3: vh2 node 4: vh6 All nodes connect to a common shared area on an iscsi device (vguests_root)
Currently vh2 and vh6 connect great, however since putting the machines back online I can no longer connect with vh1. A dmesg command on vh1 reveals the following: GFS2: fsid=: Trying to join cluster "lock_dlm", "Cluster1:vguest_roots" GFS2: fsid=Cluster1:vguest_roots.2: Joined cluster. Now mounting FS... GFS2: fsid=Cluster1:vguest_roots.2: can't mount journal #2 GFS2: fsid=Cluster1:vguest_roots.2: there are only 2 journals (0 - 1) .....
I'm having some trouble configuring clustering in a 2-node cluster, with no shared FS. Application is video streaming, so outbound traffic only...The cluster is generally ok - if I kill -9 one of the resource-applications, the failover works as expected. But it does not failover when I disconnect the power from the service owning node (simulating a hardware failure). clustat on the remaining node shows that the powered-down node has status "Offline", so it knows the node is not responding, but the remaining node does not become the owner, nor start up the cluster services/resource-applications. eth0 on each node is connected via a crossover cable for heartbeat, etc. Each eth1 connects to a switch.
I have an annoying problem with a 2 nodes cluster that I don't understand.
One of my server is totally shutdown.
When I start cman after a reboot on the other server, it hangs on "Starting fencing...". The process which is actually hanging is : /sbin/fence_tool -w -t 300 -m 45 join
When I start this program manually, I get this output :
# /sbin/fence_tool -w -t 300 -m 45 join Waiting for all 2 nodes to be members Waiting for all 2 nodes to be members Waiting for all 2 nodes to be members Waiting for all 2 nodes to be members (and so on...)
What I don't understand is why the fence_tool program wait for the 2nd node to join the cluster and don't fence the other node.
I have just installed a two server cluster with ricci luci and conga on centos 5.6 32bit , both servers are vmware guests and have a shared storage disk connected to them both
with a GFS2 file system on them + fencing agents configured to work with VMware Vcenter.
(this is supported by vmware and works great on 4 other centos clusters i have been runing for 4 monthes with no CLVMD).
In this setup i used for the first time CLVMD as recommnded by RedHat so i could have the flexablitly of LVM under the GFS2 file system but , i have been getting some Strange problem with it , some times after a developer has done some IO heavy task like unziping a file or a simple TAR the load goes to 10 - 15 and no task can be killed , trying to reboot the server hangs.
After hard shutting the server every thing works ok until the next time some one does the same IO work as before.
I just wanna simulate HPC and and other kinds of clusters in VMware workstation 7.0 , in my HP 520 laptop which is dualcore and 3Gb RAM. So can u please help me out regarding this.. I am interested to work in clusters. I am new to this hpc and other clusters..please can any one gve me document on cluster installation and configuration. I would be grateful to them.. I am using Centos OS.
Using google with search option: cman not started: Can't find local node name in cluster.conf /usr/sbin/ cman_tool: aisexec daemon didn't start.I found this URL...I have found the config_version in cluster.conf. Unfortunately, as everyone may have noticed, english is not my native tongue so I am having trouble understanding the part "Make sure you bumped the cluster config version number". Can anyone enlightened me on what should I be doing so that I could "bump" the cluster config version?
i am using the redhat cluster suite (luci and ricci) on my centos 5.3. i have 2 nodes in a cluster. when i poweroff the first node on wich a vm service is running, the service switchtes to node2. so far, so good :) but when i restart node1 the service is not failback to node1! i have created a failover domain with both nodes and priorized whre node1 has prio1 and node2 has prio2.
This is my first post, and i have a question with the fence software for VMWare ESXi. The fence_vmware agent only works with ESX, and redhat (in you GIT repository) has submited a new agent called fence_vmware_ng that claims to work with ESXi. But the problem is that they do not specify the version that works with that. Anybody has test the fence_vmware_ng agent for VMWare ESXi 4.0 ?, i follow the instructions here: [URL]...and i can install the software, the API from VMWARE site, etc, but in the moment when i run the agent nothing happen, The agent connects to server, i see in logs, but the off-reboot-on operations not works. Only works status operations, that return the state of a virtual machine. I have CentOS 5.3 (fully today updated) with RHCS.
I put together a 3 PC cluster approx. 6 years ago with Redhat, at that time I was in Grad. School. Now I am in a small company and we are planning to go for a cluster setup. I have chosen to go with Centos5.
The last time around I used NFS, RSH etc. for file sharing, communication etc. Would it be OK for me to go with the same stuff or some new software/technology has come in.
The Centos cluster setup is for high-performance computing.
I have made a cluster between two server.In luci I can see that my cluster is green and the two nodes to.I have make an IP resource and associate it to a service : green : I can relocate the service from a node to the other one and the IP appears in the list of IP addresses The problem is that I have made the same in order to configure tomcat and postgresql and it does not work...I put my configuration only for ip and tomcat:
I am researching the CentOS Cluster Setup. Does anyone know of a guide on this or have a thread linked that I can look at? I want to build a small cluster and then be able to add to it as it grows. Mainly I want to have web hosts, data, and mail behind it. The DNS will probably be on there own.