<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: redhat 6.2 cluster in Operating System - Linux</title>
    <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877849#M53969</link>
    <description>&lt;P&gt;RHEL 6.1 and newer versions have a multicast testing tool named "omping" included in the distribution.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This RHKB article includes a python script that can be used to test multicast on any system that has a python interpreter installed:&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/116553" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/116553&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;According to the latest information, using a crossover cable is possible for testing but not recommended in production:&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/151203" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/151203&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;FONT size="2" face="arial,helvetica,sans-serif"&gt;Cluster, High Availability, and GFS Deployment Best Practices:&lt;/FONT&gt;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/articles/40051" target="_blank"&gt;https://access.redhat.com/knowledge/articles/40051&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Remember that if you have a two-node cluster with no quorum disk, the vote calculation does not happen at all (two-node mode).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The cluster quorum vote calculation is very simple: if the cluster (or a part of a cluster that has lost heartbeat connection to the rest of the cluster) has more than 50% of total expected votes, the cluster has quorum and is allowed to act (i.e. accept new nodes, fence unresponsive nodes, failover services from unresponsive nodes). If there are not enough votes, there is no quorum. If there is no quorum, a cluster cannot form, and an existing cluster cannot perform any failover actions. If the part of the cluster that lost quorum was running any services, it may keep running them unless it gets fenced or gets an eviction notice through the quorum disk.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;(If the quorum is lost because all the other nodes are down, keeping the currently-running services running is the best the node can do until an admin can confirm the other nodes are in fact down. If the other nodes are OK but unreachable because of a fault in the heartbeat network, the other nodes will attempt to fence the isolated node and will take over its services if fencing is successful. If the isolated node reboots, it cannot rejoin the cluster because it has no quorum alone.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When a quorum disk is used,&amp;nbsp; it has an entirely separate logic for deciding when &amp;amp; where to grant the quorum disk votes.&lt;/P&gt;&lt;P&gt;The quorum disk daemon in each node will use the configured commands ("heuristics") to determine if the node is in a serviceable condition or not. If the node is in good condition, it will update its message slot on the shared quorum disk with a timestamp, in effect saying "I'm alive". The quorum disk daemon on the "alive" node with the smallest node ID will become the "master" quorum disk daemon: this node will grant the quorum disk votes. The extra votes will have an effect in the main quorum voting: this is the only connection between the quorum disk daemon and the main cluster quorum voting logic.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Increasing the number of qdisk votes is not really necessary in a two-node cluster: the optimum number of votes the quorum disk needs to have to achieve "last-man-standing" configuration is always the number of nodes minus 1. In a two-node cluster, this value is 1.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;(Last-man-standing: even if all the other nodes fail, the single remaining node can keep running and maintain quorum alone. It can even reboot and restart cluster services alone without manual intervention, if necessary.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;With the optimum number of quorum disk votes, you can shut down the quorum disk daemons and the cluster will still have quorum: this is useful if you need to reconfigure your quorum disk while the cluster is running. It will also make the quorum disk failure a non-critical event for the cluster.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;If you have more than the optimum number of quorum disk votes, you will lose cluster quorum if the quorum disk fails. This can make the quorum disk a single point of failure in your cluster.&lt;/P&gt;</description>
    <pubDate>Sun, 25 Nov 2012 11:15:41 GMT</pubDate>
    <dc:creator>Matti_Kurkela</dc:creator>
    <dc:date>2012-11-25T11:15:41Z</dc:date>
    <item>
      <title>redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877343#M53964</link>
      <description>&lt;P&gt;Hello Experts,&lt;BR /&gt;&lt;BR /&gt;Please help me with the exact steps on configuring two node cluster on RHEL 6.2,&lt;BR /&gt;I failed to configure the simplest cluster by below steps,&lt;BR /&gt;&lt;BR /&gt;1- install RHEL 6.2 64-bit on both nodes&lt;BR /&gt;2- add to hosts file of each server ( 1 IP in local NW and another in IP private NW).&lt;BR /&gt;x.x.x.x&amp;nbsp; node1-pub&lt;BR /&gt;z.z.z.z&amp;nbsp; node2-pub&lt;BR /&gt;y.y.y.y&amp;nbsp; node1-pvt&lt;BR /&gt;t.t.t.t&amp;nbsp; node2-pvt&lt;BR /&gt;&lt;BR /&gt;3- yum install ricci ( on both nodes )&lt;BR /&gt;4- yum install luci ( on 1 node )&lt;BR /&gt;5- yum groupinstall "high availability" ( on both nodes )&lt;BR /&gt;6- from browser, node1-pub:8084 ( login and create new cluster )&lt;BR /&gt;give cluster name, nodes name are (node1-pvt),(node2-pvt)&lt;BR /&gt;7- cluster is UP with two nodes, so far.&lt;BR /&gt;========================================================&lt;BR /&gt;Now:&lt;BR /&gt;8- configure failover domain and select both nodes.&lt;BR /&gt;9- configure resource(IP) and give IP in same range of public network.&lt;BR /&gt;10- configure servicegroup and assign the failover domain and the IP resource to the servicegroup.&lt;BR /&gt;11- IP doesn't start.&lt;BR /&gt;&lt;BR /&gt;==========&lt;BR /&gt;Nov 24 02:59:37 rgmanager start on ip "10.10.4.223/255.255.255.0" returned 1 (generic error)&lt;BR /&gt;Nov 24 02:59:37 rgmanager #68: Failed to start service:vip; return value: 1&lt;BR /&gt;Nov 24 02:59:37 rgmanager Stopping service service:vip&lt;BR /&gt;Nov 24 02:59:37 rgmanager [ip] 10.10.4.223/255.255.255.0 is not configured&lt;BR /&gt;Nov 24 02:59:37 rgmanager Service service:vip is recovering&lt;BR /&gt;Nov 24 02:59:38 rgmanager #71: Relocating failed service service:vip&lt;BR /&gt;==========&lt;BR /&gt;from luci i get this error&lt;BR /&gt;&lt;BR /&gt;Starting cluster "cluname" service "vip" from node "node1-pvt" failed: vip is in unknown state 118&lt;BR /&gt;&lt;BR /&gt;what did i miss?&lt;BR /&gt;&lt;BR /&gt;please help.&lt;BR /&gt;&lt;BR /&gt;thanks&lt;/P&gt;</description>
      <pubDate>Sat, 24 Nov 2012 08:20:51 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877343#M53964</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-24T08:20:51Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877373#M53965</link>
      <description>&lt;P&gt;Please run these commands:&lt;/P&gt;&lt;PRE&gt;# rg_test test /etc/cluster/cluster.conf start service vip&lt;BR /&gt;# rg_test test /etc/cluster/cluster.conf stop service vip&lt;/PRE&gt;&lt;P&gt;What is the output?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;According to RedHat documentation, if you use a netmask in an IP resource configuration, you should only specify the netmask length, not the fully-spelled-out netmask. In other words, use "10.10.4.223/24" instead of "10.10.4.223/255.255.255.0".&lt;/P&gt;</description>
      <pubDate>Sat, 24 Nov 2012 08:54:56 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877373#M53965</guid>
      <dc:creator>Matti_Kurkela</dc:creator>
      <dc:date>2012-11-24T08:54:56Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877545#M53966</link>
      <description>&lt;P&gt;i modified the script to add /24 instead of 255.255.255.0, i have rebooted both nodes and now the IP is up on the first node,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;below is the output of rg_test command,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;===========================================================================&lt;/P&gt;&lt;P&gt;&amp;nbsp;rg_test test /etc/cluster/cluster.conf&lt;BR /&gt;Running in test mode.&lt;BR /&gt;Loading resource rule from /usr/share/cluster/named.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/oralistener.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/postgres-8.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/ip.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/vm.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/lvm_by_vg.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/orainstance.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/samba.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/fence_scsi_check.pl&lt;BR /&gt;Loading resource rule from /usr/share/cluster/fs.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/nfsserver.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/ocf-shellfuncs&lt;BR /&gt;Loading resource rule from /usr/share/cluster/checkquorum&lt;BR /&gt;Loading resource rule from /usr/share/cluster/SAPInstance&lt;BR /&gt;Loading resource rule from /usr/share/cluster/service.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/SAPDatabase&lt;BR /&gt;Loading resource rule from /usr/share/cluster/nfsexport.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/svclib_nfslock&lt;BR /&gt;Loading resource rule from /usr/share/cluster/script.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/tomcat-6.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/clusterfs.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/openldap.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/apache.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/mysql.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/ASEHAagent.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/lvm.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/netfs.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/lvm_by_lv.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/oracledb.sh&lt;BR /&gt;Loading resource rule from /usr/share/cluster/nfsclient.sh&lt;BR /&gt;Loaded 24 resource rules&lt;BR /&gt;=== Resources List ===&lt;BR /&gt;Resource type: ip&lt;BR /&gt;Instances: 1/1&lt;BR /&gt;Agent: ip.sh&lt;BR /&gt;Attributes:&lt;BR /&gt;&amp;nbsp; address = 10.10.4.223/24 [ primary unique ]&lt;BR /&gt;&amp;nbsp; monitor_link = on&lt;BR /&gt;&amp;nbsp; nfslock [ inherit("service%nfslock") ]&lt;BR /&gt;&amp;nbsp; sleeptime = 10&lt;BR /&gt;&lt;BR /&gt;Resource type: service [INLINE]&lt;BR /&gt;Instances: 1/1&lt;BR /&gt;Agent: service.sh&lt;BR /&gt;Attributes:&lt;BR /&gt;&amp;nbsp; name = vip [ primary unique required ]&lt;BR /&gt;&amp;nbsp; domain = Active_Passive [ reconfig ]&lt;BR /&gt;&amp;nbsp; autostart = 1 [ reconfig ]&lt;BR /&gt;&amp;nbsp; exclusive = 0 [ reconfig ]&lt;BR /&gt;&amp;nbsp; nfslock = 0&lt;BR /&gt;&amp;nbsp; nfs_client_cache = 0&lt;BR /&gt;&amp;nbsp; recovery = relocate [ reconfig ]&lt;BR /&gt;&amp;nbsp; depend_mode = hard&lt;BR /&gt;&amp;nbsp; max_restarts = 0&lt;BR /&gt;&amp;nbsp; restart_expire_time = 0&lt;BR /&gt;&amp;nbsp; priority = 0&lt;BR /&gt;&lt;BR /&gt;=== Resource Tree ===&lt;BR /&gt;service (S0) {&lt;BR /&gt;&amp;nbsp; name = "vip";&lt;BR /&gt;&amp;nbsp; domain = "Active_Passive";&lt;BR /&gt;&amp;nbsp; autostart = "1";&lt;BR /&gt;&amp;nbsp; exclusive = "0";&lt;BR /&gt;&amp;nbsp; nfslock = "0";&lt;BR /&gt;&amp;nbsp; nfs_client_cache = "0";&lt;BR /&gt;&amp;nbsp; recovery = "relocate";&lt;BR /&gt;&amp;nbsp; depend_mode = "hard";&lt;BR /&gt;&amp;nbsp; max_restarts = "0";&lt;BR /&gt;&amp;nbsp; restart_expire_time = "0";&lt;BR /&gt;&amp;nbsp; priority = "0";&lt;BR /&gt;&amp;nbsp; ip (S0) {&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; address = "10.10.4.223/24";&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; monitor_link = "on";&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; nfslock = "0";&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; sleeptime = "10";&lt;BR /&gt;&amp;nbsp; }&lt;BR /&gt;}&lt;BR /&gt;=== Failover Domains ===&lt;BR /&gt;Failover domain: Active_Passive&lt;BR /&gt;Flags: none&lt;BR /&gt;&amp;nbsp; Node node1 (id 1, priority 0)&lt;BR /&gt;&amp;nbsp; Node node2 (id 2, priority 0)&lt;BR /&gt;=== Event Triggers ===&lt;BR /&gt;Event Priority Level 100:&lt;BR /&gt;&amp;nbsp; Name: Default&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; (Any event)&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp; File: /usr/share/cluster/default_event_script.sl&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;===========================================================================&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Now, i haven't configured any fencing so far, the VIP is running on the first node, now i powered it off, cluster has detected that it is off, but the vip is still on the first node which is off, and it doesn't move to first node,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;network-scripts]# clustat&lt;BR /&gt;Cluster Status for cluster @ Sat Nov 24 23:13:39 2012&lt;BR /&gt;Member Status: Quorate&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;Member Name&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; ID&amp;nbsp;&amp;nbsp; Status&lt;BR /&gt;&amp;nbsp;------ ----&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; ---- ------&lt;BR /&gt;&amp;nbsp;node1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 1 Offline&lt;BR /&gt;&amp;nbsp;node2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; 2 Online, Local, rgmanager&lt;BR /&gt;&lt;BR /&gt;&amp;nbsp;Service Name&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; Owner (Last)&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; State&lt;BR /&gt;&amp;nbsp;------- ----&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; ----- ------&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; -----&lt;BR /&gt;&amp;nbsp;service:vip&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; node1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; started&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;here is my cluster.conf&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;lt;?xml version="1.0"?&amp;gt;&lt;BR /&gt;&amp;lt;cluster config_version="10" name="cluster"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternodes&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternode name="node1" nodeid="1"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternode name="node2" nodeid="2"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/clusternodes&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;cman expected_votes="1" two_node="1"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;rm&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomains&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomain name="Active_Passive" nofailback="0" ordered="0" restricted="0"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomainnode name="node1"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomainnode name="node2"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/failoverdomain&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/failoverdomains&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;resources&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;ip address="10.10.4.223/24" monitor_link="on" sleeptime="10"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/resources&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;service domain="Active_Passive" name="vip" recovery="relocate"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;ip ref="10.10.4.223/24"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/service&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/rm&amp;gt;&lt;BR /&gt;&amp;lt;/cluster&amp;gt;&lt;BR /&gt;=======================================================&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;any clue for that? is it because that i didn't configure fencing?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;when the first node comes up, it will not join the cluster, it will just start another one considering the second node is off.&lt;/P&gt;&lt;P&gt;how to overcome this?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sat, 24 Nov 2012 17:56:16 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877545#M53966</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-24T17:56:16Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877635#M53967</link>
      <description>&lt;P&gt;The IP configuration looks good now.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;gt; Now, i haven't configured any fencing so far, the VIP is running on the first node, now i powered it off, cluster has&lt;/P&gt;&lt;P&gt;&amp;gt; detected that it is off, but the vip is still on the first node which is off, and it doesn't move to first node,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This is exactly because you have not configured fencing yet. Fencing is a mandatory part of a functioning RedHat Cluster configuration. When a node becomes unresponsive, service failover can happen only after the fencing agent confirms that the unresponsive node has been &lt;EM&gt;successfully &lt;/EM&gt;fenced. If there is no fencing configured, this step cannot complete, and the cluster cannot perform any failover operations.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;gt; when the first node comes up, it will not join the cluster, it will just start another one considering the second node is off.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This indicates a problem in cluster heartbeat delivery. Remember that the RedHat cluster heartbeat is multicast-based.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This Cisco document describes a known issue in some (many?) Cisco switches regarding multicasts and IGMP snooping, and lists five ways to fix or work around it.&lt;/P&gt;&lt;P&gt;&lt;A href="http://www.cisco.com/en/US/products/hw/switches/ps708/products_tech_note09186a008059a9df.shtml" target="_blank"&gt;http://www.cisco.com/en/US/products/hw/switches/ps708/products_tech_note09186a008059a9df.shtml&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I've seen this problem happen when the nodes are connected to different switches (i.e. located in different server rooms, separated by at least a fire-resistant wall). When both nodes join the cluster simultaneously, they both send IGMP messages to join the appropriate multicast group. The switches broadcast the IGMP messages since the multicast group is not known to them yet, so both switches become aware (by IGMP snooping) that the multicast group traffic needs to be delivered from one switch to the other. So far, so good.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;But when a node goes down, its multicast group membership times out. At that point, the switches notice that there is only one node that is interested in that multicast traffic, so the multicast traffic delivery between the switches is stopped. But when the failed node comes back up, IGMP snooping erroneously does not re-activate the multicast delivery between the switches, and as a result, each node thinks it is alone and you have a "split-brain" situation.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;With the Cisco IGMP snooping implementation, this happens when there is no source of IGMP group membership queries in the network segment. Such a source can be a real multicast router, or a "IGMP querier": a thing that sends IGMP queries exactly like a multicast router, but does not actually route any multicasts at all. It allows Cisco IGMP snooping to work correctly within a network segment.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You'll need to contact your network administrators to solve this issue. Show them the Cisco link (above) and request that they check for similar multicast behavior in the network segment that is used for your cluster heartbeat.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You'll need to get this fixed before configuring fencing: if you configure fencing while this problem is still in effect, a heartbeat network failure &lt;EM&gt;will&lt;/EM&gt; trigger a see-saw effect: node A sees B become unresponsive, so it fences node B. If power fencing is used, this causes node B to crash and reboot... at which point it will establish a new cluster, see that node A is unresponsive, and will fence node A in turn. This cycle can repeat indefinitely.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;(You could stop the see-saw effect by either setting up a quorum disk, or by configuring fencing so that a fenced node will not reboot but will stay down until manually restored instead. But these are both mitigating the symptoms instead of fixing the root cause.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have encountered this multicast issue often enough that I will always test the multicast functionality before even starting to set up a RedHat Cluster.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This RedHat Knowledge Base article is an index of cluster-related RHKB articles. Those articles contain a lot of information that is hard to find elsewhere. (RedHat Network access required)&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/articles/47987" target="_blank"&gt;https://access.redhat.com/knowledge/articles/47987&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 25 Nov 2012 00:02:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877635#M53967</guid>
      <dc:creator>Matti_Kurkela</dc:creator>
      <dc:date>2012-11-25T00:02:01Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877767#M53968</link>
      <description>&lt;P&gt;Thanks for your detailed response, that was really helping.&lt;BR /&gt;&lt;BR /&gt;Is there a tool in linux to test the multicast?&lt;BR /&gt;can i use a cross cable for heartbeat between the nodes?&lt;BR /&gt;Any best-practice for the heartbeat setup and IP configurations?&lt;BR /&gt;&lt;BR /&gt;I will use quorum with 3 votes and 1 vote for each node, could you please check the below cluster.conf? am i getting it right?&lt;BR /&gt;i am a little bit confuced with the concept redhat uses for the cluster, i am an HPUX fanatic.&lt;BR /&gt;now for the vote calculations, i made cman expected votes (5), 3 QDisk, 1 for each node = 5. how the cluster is calculation the votes to&lt;BR /&gt;become up?&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&amp;lt;?xml version="1.0"?&amp;gt;&lt;BR /&gt;&amp;lt;cluster config_version="20" name="cluster"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternodes&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternode name="node1" nodeid="1"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;clusternode name="node2" nodeid="2"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/clusternodes&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;cman expected_votes="5"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;rm&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomains&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomain name="Active_Passive" nofailback="0" ordered="0" restricted="0"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomainnode name="node1"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;failoverdomainnode name="node2"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/failoverdomain&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/failoverdomains&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;resources&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;ip address="10.10.4.223/24" monitor_link="on" sleeptime="10"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/resources&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;service domain="Active_Passive" name="vip" recovery="relocate"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;ip ref="10.10.4.223/24"/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/service&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/rm&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;quorumd device="/dev/sdb" min_score="1" votes="3"&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;heuristic program="ping -c1 -t1&amp;nbsp; "Gateway_IP" "/&amp;gt;&lt;BR /&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; &amp;lt;/quorumd&amp;gt;&lt;BR /&gt;&amp;lt;/cluster&amp;gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;/etc/hosts&lt;BR /&gt;10.10.4.221&amp;nbsp;&amp;nbsp; &amp;nbsp;server1.example.com server1&lt;BR /&gt;10.10.4.222&amp;nbsp;&amp;nbsp; &amp;nbsp;server2.example.com server2&lt;BR /&gt;172.16.26.1&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; node1.example.com node1&lt;BR /&gt;172.16.26.2&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; node2.example.com node2&lt;BR /&gt;10.10.4.223&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp; vip.example.com vip&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 25 Nov 2012 06:14:12 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877767#M53968</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-25T06:14:12Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877849#M53969</link>
      <description>&lt;P&gt;RHEL 6.1 and newer versions have a multicast testing tool named "omping" included in the distribution.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This RHKB article includes a python script that can be used to test multicast on any system that has a python interpreter installed:&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/116553" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/116553&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;According to the latest information, using a crossover cable is possible for testing but not recommended in production:&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/151203" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/151203&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;FONT size="2" face="arial,helvetica,sans-serif"&gt;Cluster, High Availability, and GFS Deployment Best Practices:&lt;/FONT&gt;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/articles/40051" target="_blank"&gt;https://access.redhat.com/knowledge/articles/40051&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Remember that if you have a two-node cluster with no quorum disk, the vote calculation does not happen at all (two-node mode).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The cluster quorum vote calculation is very simple: if the cluster (or a part of a cluster that has lost heartbeat connection to the rest of the cluster) has more than 50% of total expected votes, the cluster has quorum and is allowed to act (i.e. accept new nodes, fence unresponsive nodes, failover services from unresponsive nodes). If there are not enough votes, there is no quorum. If there is no quorum, a cluster cannot form, and an existing cluster cannot perform any failover actions. If the part of the cluster that lost quorum was running any services, it may keep running them unless it gets fenced or gets an eviction notice through the quorum disk.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;(If the quorum is lost because all the other nodes are down, keeping the currently-running services running is the best the node can do until an admin can confirm the other nodes are in fact down. If the other nodes are OK but unreachable because of a fault in the heartbeat network, the other nodes will attempt to fence the isolated node and will take over its services if fencing is successful. If the isolated node reboots, it cannot rejoin the cluster because it has no quorum alone.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When a quorum disk is used,&amp;nbsp; it has an entirely separate logic for deciding when &amp;amp; where to grant the quorum disk votes.&lt;/P&gt;&lt;P&gt;The quorum disk daemon in each node will use the configured commands ("heuristics") to determine if the node is in a serviceable condition or not. If the node is in good condition, it will update its message slot on the shared quorum disk with a timestamp, in effect saying "I'm alive". The quorum disk daemon on the "alive" node with the smallest node ID will become the "master" quorum disk daemon: this node will grant the quorum disk votes. The extra votes will have an effect in the main quorum voting: this is the only connection between the quorum disk daemon and the main cluster quorum voting logic.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Increasing the number of qdisk votes is not really necessary in a two-node cluster: the optimum number of votes the quorum disk needs to have to achieve "last-man-standing" configuration is always the number of nodes minus 1. In a two-node cluster, this value is 1.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;(Last-man-standing: even if all the other nodes fail, the single remaining node can keep running and maintain quorum alone. It can even reboot and restart cluster services alone without manual intervention, if necessary.)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;With the optimum number of quorum disk votes, you can shut down the quorum disk daemons and the cluster will still have quorum: this is useful if you need to reconfigure your quorum disk while the cluster is running. It will also make the quorum disk failure a non-critical event for the cluster.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;If you have more than the optimum number of quorum disk votes, you will lose cluster quorum if the quorum disk fails. This can make the quorum disk a single point of failure in your cluster.&lt;/P&gt;</description>
      <pubDate>Sun, 25 Nov 2012 11:15:41 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877849#M53969</guid>
      <dc:creator>Matti_Kurkela</dc:creator>
      <dc:date>2012-11-25T11:15:41Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877915#M53972</link>
      <description>Now i feel more confident, regarding the fencing, i will use brocade fencing, should i connect the brocade to the cluster network or to the server's data network?</description>
      <pubDate>Sun, 25 Nov 2012 16:49:45 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877915#M53972</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-25T16:49:45Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877925#M53973</link>
      <description>&lt;P&gt;You'll need fencing when your heartbeat fails, so putting fencing on the same network as heartbeat might not be a very good idea.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Basically do whatever is appropriate in your set-up to make sure that fencing and heartbeat connections don't fail together.&lt;/P&gt;</description>
      <pubDate>Sun, 25 Nov 2012 17:40:40 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5877925#M53973</guid>
      <dc:creator>Matti_Kurkela</dc:creator>
      <dc:date>2012-11-25T17:40:40Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5879921#M53976</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;i have configured the cluster, it is working perfectly except for the fencing,&lt;/P&gt;&lt;P&gt;i am using EMC Fabric os (DS_300B) Fabos version 6.3.2b&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;when i use the commands to fence_brocade -a IP -l user -p password -o enable/disable it returns successfull, but i doesn't really work when the cluster node is fencing the other.&lt;/P&gt;&lt;P&gt;fence_tool ls returns fencing and i have to fence_ack_manual to move the services,&lt;/P&gt;&lt;P&gt;then fence_node -U&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Do i have to do something for that?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Tue, 27 Nov 2012 03:32:58 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5879921#M53976</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-27T03:32:58Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5879927#M53977</link>
      <description># fence_brocade -a 192.168.201.167 -l admin -p password -n 2 -o enable&lt;BR /&gt;success: portenable 2&lt;BR /&gt;# fence_brocade -a 192.168.201.168 -l admin -p password -n 2 -o enable&lt;BR /&gt;success: portenable 2&lt;BR /&gt;&lt;BR /&gt;#fence_ack_manual node1&lt;BR /&gt;About to override fencing for node1.&lt;BR /&gt;Improper use of this command can cause severe file system damage.&lt;BR /&gt;&lt;BR /&gt;Continue [NO/absolutely]? absolutely&lt;BR /&gt;Done&lt;BR /&gt;&lt;BR /&gt;#grep brocade /etc/cluster/cluster.conf&lt;BR /&gt;&amp;lt;fencedevice agent="fence_brocade" ipaddr="192.168.201.167" login="admin" name="Brocade1" passwd="password"/&amp;gt;&lt;BR /&gt;&amp;lt;fencedevice agent="fence_brocade" ipaddr="192.168.201.168" login="admin" name="Brocade2" passwd="password"/&amp;gt;&lt;BR /&gt;Thanks</description>
      <pubDate>Tue, 27 Nov 2012 03:39:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5879927#M53977</guid>
      <dc:creator>karmellove</dc:creator>
      <dc:date>2012-11-27T03:39:14Z</dc:date>
    </item>
    <item>
      <title>Re: redhat 6.2 cluster</title>
      <link>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5881083#M53983</link>
      <description>&lt;P&gt;You should read these if you use storage-based fencing like fence_brocade:&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/220323" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/220323&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://access.redhat.com/knowledge/solutions/236483" target="_blank"&gt;https://access.redhat.com/knowledge/solutions/236483&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In a nutshell: while storage-based fencing certainly stops the fenced node from writing to the disks, the fenced node won't know about it immediately. Only after the fenced node makes an I/O request and sees it time out (after multiple retries) , the node becomes aware that it has been fenced, and can stop its network activity. Depending on multipath timeout and queue_if_no_path settings, the I/O requests may take a long time to time out - or with queue_if_no_path, they might never time out at all. You may need to tune the I/O request timeouts to be compatible with cluster timeouts.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Another issue is that when a node is fenced with a storage-based method, the fenced node will quite likely contain data in its buffers that has not been written to the shared disks - which are now inaccessible because of fencing. Normally, Linux kernel tries very very hard to not lose data, so even "reboot -f" will attempt to write out the buffers... so in the case of a SCSI-based fencing, the system will fail to reboot.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In this specific situation, you actually *want* to lose the content of the buffers, since the data may already been made obsolete by the other node taking over the service. You should never unfence a cluster node that has been fenced by a storage-based method without first making sure the node has no longer any data destined to the shared disks in its buffers. A hard reboot by power-cycling is the surest way to do this.&lt;/P&gt;</description>
      <pubDate>Tue, 27 Nov 2012 21:21:41 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/redhat-6-2-cluster/m-p/5881083#M53983</guid>
      <dc:creator>Matti_Kurkela</dc:creator>
      <dc:date>2012-11-27T21:21:41Z</dc:date>
    </item>
  </channel>
</rss>

