So far, I can ping a virtual ip, and manually relocate it between the nodes, but I didn't figure out, how to do this automatically. So this is my question: How can I setup the cluster, to it automatically failover the a service to another node case one node fails?
I want to automate the following manual process.Currently, I am encryptying a set of files using openssl as follows:Encrypt file.txt to file.out using 256-bit AES in CBC mode $ openssl enc -aes-256-cbc -salt -in file1 -out file1.enI am then prompted for a password, which is then used to encrypt the fileWhen decrypting, I type $ openssl enc -d -aes-256-cbc -infile1.enc -out fileI am then prompted for the password - which again, I manually type.I want to automate this process of en/decryption - so I need to find a way of providing openssh with the password.
My first thought is whether it is possible to read the password from a file (say)? Or is there a better way of doing this?Also, I suppose that I will have to place restriction on who can view the password file - otherwise, that defeats the whole objective of using a password. I am thinking to run the bash script as a specific user, and then give only that user read rights to the contents of that file.Is this the way its done - or is there a better way?Ofcourse all of this leads to yet another question - which is, how to run a bash script as another user - without having to type the user pwd at the terminal.
I have installed dhcp-server on ubuntu. And server is providing ip address to clients (Window machines) but not default-gateway but I have another network in my network and same configuration and same dhcp-server provides every information.
I have created a simple menu driven script for our Operations to take care of the basic monitoring and managing of our production application from the back-end. Now, the script when tested in UAT environment was fine, but when deployed to production it kind of behaved oddly.hen the Operator chooses an option from the menu he is given the output and at the end is prompted to return to the main menu by using ctrl+c. In production, this return does not occur for some strange reason and the program just sits there.The session becomes unresponsive after that and I'm forced to terminated it by closing the PuTTY.I tried enabling the debug mode too (set -x) and still was not able to find any useful hints/trails as to why.
I don't have much experience in clustering. And I'm deploying a cluster system on CentOS.But I don't know how long a node failover and another node take over those resouces to continue running service is good, fast or slow? 1s, 10s or
I have lack of understanding of CentOS in general. I have looked for a remedy on other forums and google, but haven't been able to find the answer. I have a 3 node cluster that was functioning great until I decided to go offline for awhile. My config is as follows: node 2: vh1 node 3: vh2 node 4: vh6 All nodes connect to a common shared area on an iscsi device (vguests_root)
Currently vh2 and vh6 connect great, however since putting the machines back online I can no longer connect with vh1. A dmesg command on vh1 reveals the following: GFS2: fsid=: Trying to join cluster "lock_dlm", "Cluster1:vguest_roots" GFS2: fsid=Cluster1:vguest_roots.2: Joined cluster. Now mounting FS... GFS2: fsid=Cluster1:vguest_roots.2: can't mount journal #2 GFS2: fsid=Cluster1:vguest_roots.2: there are only 2 journals (0 - 1) .....
I'm having some trouble configuring clustering in a 2-node cluster, with no shared FS. Application is video streaming, so outbound traffic only...The cluster is generally ok - if I kill -9 one of the resource-applications, the failover works as expected. But it does not failover when I disconnect the power from the service owning node (simulating a hardware failure). clustat on the remaining node shows that the powered-down node has status "Offline", so it knows the node is not responding, but the remaining node does not become the owner, nor start up the cluster services/resource-applications. eth0 on each node is connected via a crossover cable for heartbeat, etc. Each eth1 connects to a switch.
Using google with search option: cman not started: Can't find local node name in cluster.conf /usr/sbin/ cman_tool: aisexec daemon didn't start.I found this URL...I have found the config_version in cluster.conf. Unfortunately, as everyone may have noticed, english is not my native tongue so I am having trouble understanding the part "Make sure you bumped the cluster config version number". Can anyone enlightened me on what should I be doing so that I could "bump" the cluster config version?
I am working on the beginning of implementing a two-node cluster with shared storage (GFS) and IP address. Both machines are virtual on VMware ESX 3.5, that should not make a difference, but that is the background.current status is that I have a single node cluster built with only the IP address configured within the cluster. The issue that I am having is that I have configured a service to contain only the IP address resource, however, when I go into cluster management that "service" does not register. As such, I cannot bring it online, ping it, etc. below is my cluster.conf configuration:
I've configured drbd with Heartbeat. The nodes are connected at first, but when I issue "/usr/lib/heartbeat/hb_standby" on node 1, node 2 won't take over, and after a while, both nodes become "WFConnection".
I figure this happens because I created "dev/drbd0" on both nodes! Do you know how to remove it? I googled it but got nothing.
i have a ubuntu server and no of nodes.give me any suggesions for how to handle the node users login with my permission?.i mean than they can switching on and boot the system but the system want to stay in unusable stage.After getting my permission they can operate the system.
I have just started a job as an IT tech, and have inherited a linux server running CEntOS 5.3 -- I have some experiance with linux, but am a little at sea. I'm using Webmin 1.490.The main use of the server is to samba share file space to windows clients, running XP.Everyone can happily connect to their home share, however the other shares have problems. One of the shares "staff" fails intermittently (to some staff members) and some of the other shares fail always.
The failure is that when a user tries to connect to the share, it pops up a username and password dialog (even if they just successfully viewed their home dir) then on entering correct details and clicking ok, the box refreshes with <domain><user name> (where braced values are replaced by literal values). If a user is connecting from outside via vpn (hosted though a different department's server) the domain is replaced by their computer name.If the password is re-entered, the same happens, in an infinite loop, and no access is ever given.When I say the staff share is intermittent, there likely is some pattern, but I haven't been able to work it out yet -- I think it may have something to do with how recently the account was set up, or physical location.
I can not get the node or cloud controllers to startup using the init.d scripts. I have a fresh install of CentOS 5.4 with Eucalyptus 1.6.2 I have compiled Eucalyptus and all packages using the RPM supplied from Eucalyptus and utilizing yum installer. I do not currently have any processes or applications listening currently on the ports on the boxes as well. I think it may be a permissions issue or something because I get a "permission denied error", but I am not sure if it is Eucalyptus or CentOS. It looks as if it is not binding to the address on the interface of the NIC. It may be something else however. I have the Node controller, Cloud controller, and Cluster controller on seperate physical boxes. When I try to run either the cloud controller or the node controller I get this message:
[root@cluster-cont ~]# /etc/init.d/eucalyptus-cc start Starting Eucalyptus cluster controller: (13)Permission denied: make_sock: could not bind to address [::]:8774 (13)Permission denied: make_sock: could not bind to address 0.0.0.0:8774
My requirement is to route ssh sessions from a single head node to multiple slave nodes. So what i want is, for a client there is just one point of entry (master/head node) to ssh into, it evaluates the load on the slave nodes connected on to internal network and routes the ssh session, kind of a ssh load balancer. Do you have any idea what open source solution i can apply for my problem?
I have tried using LVS piranha, it works well for http and https load balancing but not for ssh load balancing.
I have two machines between which I need to share a folder.On server1, I have the user 'appuser' that needs to access (read/write/delete) on this share.On server2, 'root' accesses this share and writes to it.I have the following in /etc/exports on server1:/home/app-share 999.999.99.99/28(rw,insecure,sync,no_root_squash)where the number is the IP address. How can I change this to allow 'appuser' access?
I am using Linpus lite (a fedora based linux) on an ACER ONE notebook. I use a GUI to install/update or remove software. It is called pirut. I believe it a GUI of the yum command. The repository this install program uses is the fedora one. I find the following description of the repository in the menus of that GUI: Name: Fedora Description:Fedora 8 - i386 Mirror list : h[URL]... The problem is the following: The versions I can install from that GUI are outdate. For example, the firefox I have is 184.108.40.206 although the newest one available at the firefox website is 3.6.8. The same with openoffice, gcc and many many other tools.
This is embarrassing: I wanted recently to install a software on my own which was not available at that repository. This software requires recent versions of many tools. It's embarassing to have to update them manually..
I assumed the default repository of fedora should have access to the newest versions. If that is true, then the problem is in my system and not in the repositories. In that case, does any one have a clue on how to fix that problem? Now if the repository I use is indeed outdated, Should I change to another repository? Can anyone indicate which repository is compatible with my system and how to add it?
I've been looking around the net for executing a shell script. My basic understanding is that after setting executable permissions and providing a path (#!/bin/sh) in the first line of the script, I can type ./myscript to execute instead of sh myscript. This is not working for me. I can run "sh myscript" but not "./myscript" even though I know for sure I have across the board execute permission and my sh path is correct. I'm working on a redhat linux station.
I currently run Win 7 and want to upgrade my computer to a server to accomplish the following... I have a VPN Service. I want the server machine to connect to a VPN providing a secure connection. Then, I want all the machines in the house (windows based) to connect through the server onto the VPN connection. Hopefully this makes sense. Would it be better to stick to Windows Server 2008 or switch to Ubuntu?
How can I figure out the "Device" name from the "Sysfs" node? e.g. [root@baba 0000:00:01.0]# lspci | grep -i Ethernet 00:19.0 Ethernet controller: Intel Corporation 82566DC Gigabit Network Connection (rev 02)
hello i am trying to change my password, but when i type in the new password i get this:"The password is longer than 8 characters. On some systems, this can cause problems. You can truncate the password to 8 characters, or leave it as it is."my question is what kind of problem could i get and how can i change so i have to log in every time i start the computer?