<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: After a reboot in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517232#M673834</link>
    <description>If 2 of 3 nodes goes down unexpectedly, the 3rd node would not continue running Serviceguard as dictated by cluster reformation protocol.&lt;BR /&gt;See page 117 in the latest "Managing Serviceguard" manual: &lt;A href="http://docs.hp.com/en/B3936-90143/B3936-90143.pdf" target="_blank"&gt;http://docs.hp.com/en/B3936-90143/B3936-90143.pdf&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;The rule of thumb is:&lt;BR /&gt;&amp;gt;50%&lt;BR /&gt;If, after a sudden loss of HB connection to other nodes, the remaining nodes are a majority subset of the original cluster, they automatically reform a cluster, continue operations of currently running  packages on these node, and adopt dead node packages.&lt;BR /&gt;&lt;BR /&gt;=50%&lt;BR /&gt;If, after a HB outage, an even split occurs between active nodes, arbitration in the form of quorum server or cluster lock disk must be sought to receive permission (or denial) to reform a cluster.  The first side to contact the arbitrator reforms the cluster.  The last side to reach the arbitration device must TOC/reboot to insure data integrity.&lt;BR /&gt;&lt;BR /&gt;&amp;lt;50% (your case)&lt;BR /&gt;If, after a HB outage, remaining node(s) find themselves in a minority subset of the original cluster, must TOC/reboot to insure data integrity.  The assumption here, is that a majority of nodes survived the HB failure and will take control.  This is necessitated by a choice as to how to deal with a uneven split between active nodes.&lt;BR /&gt;&lt;BR /&gt;--------&lt;BR /&gt;If the scenario you describe however, is the result of graceful exits (cmhaltnode) by 2 nodes in sequence wherein the remaining node was the sole operator, normal package failover protocol dictates whether package failover to the last node would occur.&lt;BR /&gt;&lt;BR /&gt;Remember that each package has been configured with either a list of adoptive nodes or a * in the NODE_NAME parameter.  If a list, the adoptive node list may be a subset of all nodes, and the list dictates failover order in the event of a node departure.  &lt;BR /&gt;If a *, any node may be an adoptive node for the package.  &lt;BR /&gt;If FAILOVER_POLICY is configured to MIN_PACKAGE_NODE, the package is moved to the remaining node that has the fewest packages on it.  &lt;BR /&gt;&lt;BR /&gt;Ultimately, if 2 nodes leave the cluster gracefully and package AUTO_RUN and node_switching are enabled for their packages, the packages will halt on the departing node(s) and will be started on the last node.&lt;BR /&gt;&lt;BR /&gt;+NOTE: upper-case package parameters denote the package was configured with legacy format.  Change to lower-case for modular package format.</description>
    <pubDate>Tue, 20 Oct 2009 10:24:56 GMT</pubDate>
    <dc:creator>Stephen Doud</dc:creator>
    <dc:date>2009-10-20T10:24:56Z</dc:date>
    <item>
      <title>After a reboot</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517231#M673833</link>
      <description>In a three node cluster...two nodes go down. All the pkgs would hence failover on the third node.&lt;BR /&gt;&lt;BR /&gt;In case the third node reboots, would all the packages start on the third node? Acc to me, yes they should.&lt;BR /&gt;&lt;BR /&gt;Hope quorum would not come in picture here</description>
      <pubDate>Tue, 20 Oct 2009 09:17:42 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517231#M673833</guid>
      <dc:creator>Buds</dc:creator>
      <dc:date>2009-10-20T09:17:42Z</dc:date>
    </item>
    <item>
      <title>Re: After a reboot</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517232#M673834</link>
      <description>If 2 of 3 nodes goes down unexpectedly, the 3rd node would not continue running Serviceguard as dictated by cluster reformation protocol.&lt;BR /&gt;See page 117 in the latest "Managing Serviceguard" manual: &lt;A href="http://docs.hp.com/en/B3936-90143/B3936-90143.pdf" target="_blank"&gt;http://docs.hp.com/en/B3936-90143/B3936-90143.pdf&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;The rule of thumb is:&lt;BR /&gt;&amp;gt;50%&lt;BR /&gt;If, after a sudden loss of HB connection to other nodes, the remaining nodes are a majority subset of the original cluster, they automatically reform a cluster, continue operations of currently running  packages on these node, and adopt dead node packages.&lt;BR /&gt;&lt;BR /&gt;=50%&lt;BR /&gt;If, after a HB outage, an even split occurs between active nodes, arbitration in the form of quorum server or cluster lock disk must be sought to receive permission (or denial) to reform a cluster.  The first side to contact the arbitrator reforms the cluster.  The last side to reach the arbitration device must TOC/reboot to insure data integrity.&lt;BR /&gt;&lt;BR /&gt;&amp;lt;50% (your case)&lt;BR /&gt;If, after a HB outage, remaining node(s) find themselves in a minority subset of the original cluster, must TOC/reboot to insure data integrity.  The assumption here, is that a majority of nodes survived the HB failure and will take control.  This is necessitated by a choice as to how to deal with a uneven split between active nodes.&lt;BR /&gt;&lt;BR /&gt;--------&lt;BR /&gt;If the scenario you describe however, is the result of graceful exits (cmhaltnode) by 2 nodes in sequence wherein the remaining node was the sole operator, normal package failover protocol dictates whether package failover to the last node would occur.&lt;BR /&gt;&lt;BR /&gt;Remember that each package has been configured with either a list of adoptive nodes or a * in the NODE_NAME parameter.  If a list, the adoptive node list may be a subset of all nodes, and the list dictates failover order in the event of a node departure.  &lt;BR /&gt;If a *, any node may be an adoptive node for the package.  &lt;BR /&gt;If FAILOVER_POLICY is configured to MIN_PACKAGE_NODE, the package is moved to the remaining node that has the fewest packages on it.  &lt;BR /&gt;&lt;BR /&gt;Ultimately, if 2 nodes leave the cluster gracefully and package AUTO_RUN and node_switching are enabled for their packages, the packages will halt on the departing node(s) and will be started on the last node.&lt;BR /&gt;&lt;BR /&gt;+NOTE: upper-case package parameters denote the package was configured with legacy format.  Change to lower-case for modular package format.</description>
      <pubDate>Tue, 20 Oct 2009 10:24:56 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517232#M673834</guid>
      <dc:creator>Stephen Doud</dc:creator>
      <dc:date>2009-10-20T10:24:56Z</dc:date>
    </item>
    <item>
      <title>Re: After a reboot</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517233#M673835</link>
      <description>Considering the policies to be automatic failover=yes and configured node= 3rd node all packages would failover to the third node.&lt;BR /&gt;Now, in case the third node also reboots, the first two still being down, would all the packages start on the third node?</description>
      <pubDate>Wed, 21 Oct 2009 11:11:20 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517233#M673835</guid>
      <dc:creator>Buds</dc:creator>
      <dc:date>2009-10-21T11:11:20Z</dc:date>
    </item>
    <item>
      <title>Re: After a reboot</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517234#M673836</link>
      <description>After rebooting, a server may run '/sbin/init.d/cmcluster start' if the /etc/rc.config.d/cmcluster contains AUTOSTART_CMCLD=1.&lt;BR /&gt;The cmcluster script eventually performs cmrunnode.  That command has 2 modes of operation:&lt;BR /&gt;1) If partner nodes are reachable and running Serviceguard, join that cluster&lt;BR /&gt;2) If partner nodes are reachable but not running a cluster, start cluster formation stage, waiting for -ALL- other nodes to "vote" to start the cluster.  This wait stage lasts for AUTOSTART_TIMEOUT period (default 10 minutes as identified in the cluster config file) before terminating with node cluster daemons running. &lt;BR /&gt;&lt;BR /&gt;In your scenario of 2 nodes down, the 3rd  node running boot-time scripts will, upon running /sbin/rc3.d/S800cmcluster attempt to form a cluster, wait 10 minutes and cease attempting to form a cluster with it's peers.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;The reason being that this one node cannot know the status of the other cluster member nodes - they may be running but the rebooting node may not be able to contact them due to a local heartbeat network NIC failure.&lt;BR /&gt;</description>
      <pubDate>Thu, 22 Oct 2009 11:20:39 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/after-a-reboot/m-p/4517234#M673836</guid>
      <dc:creator>Stephen Doud</dc:creator>
      <dc:date>2009-10-22T11:20:39Z</dc:date>
    </item>
  </channel>
</rss>

