I need to build a 3 node web server cluster to run a php application. Since the app requires users to login (which means a session state is to be maintained), I will be sharing sessions save path, I also need to share the application directory across 3 nodes. I having trouble deciding which cluster file system to select.
I have problem with port based routing for local traffic. I can't use trick with iptables -t mangle, ip route table 1, ip rule fwmark table 1 because it works only with forwarded packets. I can't even use patch-o-matic because it's obsolete. And xtables-addons doesn't contain support for "-j ROUTE" yet.
I'm trying to build a Fedora 12 cluster to run DHCP on top of two servers using pacemaker and drbd. The pacemaker runs great but i'm having the hardest time getting DRBD installed.
My kernel version is 22.214.171.124-99.fc12.i686.PAE First i tried to [root@dhcp-primary drbd]#yum install drbd this doesn't give me everything because apparently the kernel module for DRBD must be built from source. so i unpacked drbd-8.3.6.tar.gz and tried :
I'm not very experienced in compiling from source i'd love to make an rpm for this but after scouring the web and trying different install methods, I am still at a loss.
I'd like a way to see all of the devices on my local network and what their local IP address is. I recall that I used wireshark to troubleshoot a similar problem a while back, but it doesn't seem to have a way to see all of the devices- only the traffic. (I'd like to do this without having to physically interface with my router if possible, and I am in an encrypted network if that matters)
I have installed a web server on my local network. Everything is well configured and web pages are shown correctly from Internet (outside the local network) using the domain or the public IP.The issue is if I try to see that web pages (using the domain or the public IP) from inside the local network. In that case the router config page (192.168.1.1) is shown instead of the web pages.From inside the local network I'm only able to see the web pages using the internal IP address (192.168.1.XX).
I want to write a custom rule to allow all connections to the ip addresses on my local network (192.168.2.2 through ...99) but I don't know how. I know adding a custom rule asks me to read a file and put it in "iptables" format, but I don't know how...
I am using Centos. I have read places that you can use Drbd + heartbeat + nfs to make a simple failover NFS server.I can't find any document that works though. I've tried 20 or so, including some Debian ones.So, does anyone have any other ideas of how to do this?Point me in the right direction please.I want 2 nodes. One to be actively serving an NFS share The other to be ready for failover. If the first one goes out, the second takes over.Meaning, the filesystem is in sync, the IP must change, and NFS must come up
I've got an Ubuntu server hosting our websites and other various things here in our own home. We recently switched to a router that doesn't support loopback (abomination), so I've set up hosts files on our computers so we can access our own sites when on our home LAN.
However, we often take our laptops as we travel about, and I'm guessing due to the hosts files when we try to access our sites, it'll look on whatever local network we're connected to for our server, which won't work, obviously.
Is there a way to set up something like a hosts file that'll only try to look up the local IP of the server when we're on a specific network (our home one), or have one that tries to look for the local IP first, then proceeds to try and resolve the domain name and use the external IP if the local IP doesn't work?
I need to setup an linux cluster ..so i prefer ubuntu because of support and i personally i use ubuntu.. and can any one explain in breif ..what all the things needed to setup an ubuntu based cluster my configuration for each node will be (totally 6 nodes) core2 duo with 4 gb ram i need 4 nodes and 2 for load balancing..
I'm looking for a write one, read many file synchronization, file system or similar technique. In fact I want a to imitate a zfs_send/zfs_receive similar behaviour on Linux. My requirements are pure local (or in terms of distributed file systems: rack-aware) file handling with remote synchronization. Some random features/requirements I want to achieve are (I use the terms file and "file system" here interchangeable, as I don't care wheter I have to synchronize file based or use full file system supporting that): Access on files from a specific host is pure local, i.e. the node has a full copy of a file, no parts are distributed. Read and write access is local, the remote side gets synchronized with write changes only Write changes on the remote side are non-blocking and asynchronous No need for concurrent file access, concurrent, distributed locking and so on Every file is changed by a single node only, other nodes have read access The file "owner", thus the node allowed to write files must be changeable Similar techniques (some as technology concept, not usable for me) exist, among those DRBD, MySQL Replication or even pure rsync.
I don't care whether I'm referring to a pure local solution (i.e. between two block devices) or a client/server mode. DRBD won't fit because its a peer to peer solution, I need to synchronize a lot of clients with a central storage through that way where DRBD just mirrors from peer to peer among two nodes and doesn't allow another node to take over the role of the primary node (i.e. node A is master, storage B secondary, I want to change to a thrid node C which becomes master and receives hence a full update from storage B.
Just as a guess, I'd make a local LVM volume on every node where write access happens. In the same time there exists a iSCSI target with multipath to a remote hosts which is shared with every node, again providing a clustered LVM. So I'd need to synchronize a local LVM volume with a remote LVM image. The question is just by which technique I could achieve that (i.e. as RAID 0, where every read and write goes to one node only and doesn't wait the other node to succeed). I could achieve this with rsync by a cyclic push to the remote host, but the problem is, that rsync isn't block based thus I'd need to synchronize every time the whole file, not only changes.
I am trying to build GFS2 cluster with 2 or 3 Fedora 14 nodes, but I've encountered some problems from the start. First luci does not work at all in Fedora 14. There is no luci_admin and even if I manage to start luci service, I get a blank white screen when I try to open it from the browser. I've googled a bit and found that I'd might be able to setup GFS if I manage to build cluster.conf manually and start the cluster suite, but I cannot find documentation on how to create cluster.conf anywhere. If anyone knows how to setup GFS2 without a cluster suite or how to configure cluster.conf.
Using google with search option: cman not started: Can't find local node name in cluster.conf /usr/sbin/ cman_tool: aisexec daemon didn't start.I found this URL...I have found the config_version in cluster.conf. Unfortunately, as everyone may have noticed, english is not my native tongue so I am having trouble understanding the part "Make sure you bumped the cluster config version number". Can anyone enlightened me on what should I be doing so that I could "bump" the cluster config version?
I just installed open Suse 11.3, and I cannot SSH my school. Upon further investigation I could not even ping any machines outside my local area network. Ironically I could nmap machines outside my local area network.
I have a mac and an openSUSE box connected wirelessly to a router. From my mac, if I ping the Linux machine by both ip address and hostname "elmo.local", I get a response. If I ping my mac from the Linux machine using its ip address, I get a response. However, if I ping the mac from the Linux box using its name "kermit.local" - nothing!
This means in order to share files between them via the network, I have to use the ip address rather than name, but I'd rather not. When I had Ubuntu installed instead of openSUSE this all worked fine out of the box, so I'm assuming it's a problem with the setup of openSUSE rather than the router or the mac.
I'm working with a Ubuntu 10.04 LTS system with two network interfaces (both Ethernet). I wish to setup this system such that it is simultaneously connected to my local and an OpenVPN network and able direct traffic between the connections depending on what program is sending the traffic. The problem: Under my current OpenVPN configuration all network traffic is directed to the VPN.
In practice, I would like OpenVPN to operate out of one of my two network interfaces and leave the other interface connected to the local network. Then by default all network traffic should be directed to my local network unless I specify (on a per program bases) that certain traffic should go though the VPN. These two network connections can (should) stay completely independent of each other and do not need to talk to each other.
I have two networks. One of them is wired, the other is wireless. The wired has an internet connection and a few other computers connected to it. The wireless network has a few hosts connected to it too, but it has no internet connection. What I've been trying, fruitlessly, to do, is make all connections that are bound to the internet, or my wired network, be routed that way, and all the connections to the hosts of the wireless network go that way.
Here's the setup..
192.168.1.0/24 Gateway = 192.168.1.1
Route internet through here
192.168.2.0/24 Gateway = 192.168.2.1
If my computer sends a packet to the internet, it should be routed through 192.168.1.1 If I send a packet to one of the local hosts of the wireless network, it should be routed through 192.168.2.1. Here's the routing table I've set up(This is one of many configurations I've tried)
$ ip route show 192.168.2.1 dev wlan0 scope link 192.168.2.0/24 via 192.168.2.1 dev wlan0 src 192.168.2.4
With this, and everything else, I get destination host unreachable when pinging. The strange thing is that, if I unplug my eth cable, reboot and connect to the wireless network, everything is fine and I can access the router and the others. I'm trying to improve my networking skills, as I've had this of setting up a small linux box as a router for quite some time, for the fun of it, but I need to get routing under control before I go ruin my network.
I've been running Fedora on my Laptop (Dell Precision) for 1.5 years. I have a Windows based network. What I'd like to do now is to change that network to a Linux (Fedora) based network. Its got 4 workstations and a switch ... I don't want internet access to this network, ... just to be able to share files between the PCs. In windows it was a matter of specifying different IPs and same workgroup assignments to each node. How does it work in Fedora ?I'd also like to see Peer-to-Peer work between my Laptop and my friend's Acer laptop running (Fedora too).
i do love to learn this stuff so i thing I'll give it a shot as far as setting mail, ftp, and other services i think i can manage. what i really has no clue of with where to begin is how to make this: there are foot computers, 2 of them need to access the net and 2 should not. of course, all of the PC's should be able to talk to the main server (the ftp, mail , file sharing server)so, where is my starting point?
I'm trying to get experience in creation of virtual Fedora 9 x64 cluster using Ubuntu 9.04/Virtual Box 2.2.2. I have a head node with 1 G of RAM and two slave nodes with 512 M each. There's internal networking in Virtual Box with DHCP, so my nodes have addresses:
192.168.15.100/24 (Slave0.localdomain), 192.168.15.101/24 (Slave1.localdomain), 192.168.15.102/24 (Head.localdomain). Ping is OK. I've added the following string to /etc/exports file: /usr/share 192.168.15.0/24(rw) And then used /usr/sbin/exportfs -a But showmount -e 192.168.15.102 entered at slave node shows nothing. What's wrong? May be I have to somehow start nfsserver, but how?
At home we have a small home network, but it is very slow. I believe it is, because some computers are placed in a workgroup, called WORKGROUP and others are placed in a domain called TUE. The domain is because many of the laptops at our house also have to work on our university, and I can't simply put them in a workgroup.
The reason why I believe it is slow because of this, is because file sharing between the computers in the workgroup goes with 5 MB/s. File sharing between a domain computer and a workgroup computer goes as fast as 160 KB/s. This is the upload speed of our internet here. How to make the connection with the domain computers the same speed as the workgroup computers? I can not simply put the domain computers out of their domain and in the workgroup (as explained before)
I setup a server with vmware esxi 4. I installed opensuse 11.2 and I have successfully setup an internet gateway and I manually configured (static) my client IP. My question is, is there any software or commands that I could use to manage the bandwidth usage for each client? eg. only 5Mbps for 192.168.1.5, 10Mbps for 192.168.1.10, etc
Is there a tool or tutorial to build a distro based on mint/ubuntu/debian?Git, scripts, and tools, small compilations for branding are OK. Compiling kernel, x11, gnome, compiz, etc. is BAD unless really required. Specifically, I'm looking for a way to have some applications installed by default, change logos/about boxes, change theme, configure what compiz options are, and add firmware drivers (connect to the Internet to download wireless adapter driver? <- headache, gotta find a lan cable now ...), and still have the liveCD install method. I'd like it to be based on linux mint, but I only want a few tools from it (update manager, software manager, flash, media codecs), so it's ok if it's ubuntu/debian with those mint tools added back in.
I saw some options, too complex, and others too simple. Not looking for a "learning experience" like LFS which gives me a horrible linux build if I don't do something exactly right. Nor any of those tools which are just package pickers and don't do enough. suse studio looked about right(maybe tad too easy), but was RPM based, not deb based.
I have a Linksys router configured via Windows 7 and have to other PC's using Ubunto, and would like to share folders on the 2 Linux boxes with the Windows Laptop. I can see the windows network through the Linux machines but not vice-versa. I also have a Konica 1400W printer connected to the Linux box but can't get the W7 laptop to find it.Sharing message box states I haven't installed the correct sharing packages, but cant find them