<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Desaster Test in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599463#M716014</link>
    <description>I could not retrieve your attachment, but from your description, yes it is normal, although you didn't need to perform a manual TOC on the primarys to deactivate the vgs.  When you experience a failure of some kind (i.e. network), the "heartbeat" that is sent between the two nodes stops, which causes each node to race for the cluster lock disk (required in two node clusters).  The node that gets the lock disk forms a one node cluster, the other panics.  when you perfom a cmruncl, the following happens with respect to your applications:&lt;BR /&gt;&lt;BR /&gt;1, volume group activation&lt;BR /&gt;2. check and mount file systems&lt;BR /&gt;3. assign pkg ip&lt;BR /&gt;4. start user  defined run commands&lt;BR /&gt;5. start service processes&lt;BR /&gt;&lt;BR /&gt;which is why your oracle and sap were started automatically after cmruncl (they are part of #4)&lt;BR /&gt;&lt;BR /&gt;Hope this helps&lt;BR /&gt;Chris</description>
    <pubDate>Tue, 23 Oct 2001 09:19:51 GMT</pubDate>
    <dc:creator>Christopher McCray_1</dc:creator>
    <dc:date>2001-10-23T09:19:51Z</dc:date>
    <item>
      <title>Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599462#M716013</link>
      <description>Dear experts,&lt;BR /&gt;We have several 2-node Clusters in 2 different datacenters runnig SAP with the MC/SG SAP extension.&lt;BR /&gt;We recently have performed a desaster test to check if things behave as they should. For this thest we have cut all lines (Network, Fibre-channel...) to simulate the loss of the entire Datacenter.&lt;BR /&gt;The primary nodes with the oracle DB's are running in the datacenter we shut down.&lt;BR /&gt;Result: the alternate nodes have TOC'ed, the primary nodes remained up, and could not remove the volume groups, until we have done a manual TOC. After reboot of the alternate nodes and cmruncl the cluster came up (asked to make shure the primary nodes are really down) and Oracle/SAP was started.&lt;BR /&gt;NODE_TIMEOUT=6000000&lt;BR /&gt;HEARTBEAT_INTERVAL=2000000&lt;BR /&gt;NODE_FAIL_FAST_ENABLED=yes&lt;BR /&gt;Is this behaviour correct?&lt;BR /&gt;(complete config attatched)&lt;BR /&gt;Thanx&lt;BR /&gt;</description>
      <pubDate>Tue, 23 Oct 2001 09:07:40 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599462#M716013</guid>
      <dc:creator>Paul Barmettler</dc:creator>
      <dc:date>2001-10-23T09:07:40Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599463#M716014</link>
      <description>I could not retrieve your attachment, but from your description, yes it is normal, although you didn't need to perform a manual TOC on the primarys to deactivate the vgs.  When you experience a failure of some kind (i.e. network), the "heartbeat" that is sent between the two nodes stops, which causes each node to race for the cluster lock disk (required in two node clusters).  The node that gets the lock disk forms a one node cluster, the other panics.  when you perfom a cmruncl, the following happens with respect to your applications:&lt;BR /&gt;&lt;BR /&gt;1, volume group activation&lt;BR /&gt;2. check and mount file systems&lt;BR /&gt;3. assign pkg ip&lt;BR /&gt;4. start user  defined run commands&lt;BR /&gt;5. start service processes&lt;BR /&gt;&lt;BR /&gt;which is why your oracle and sap were started automatically after cmruncl (they are part of #4)&lt;BR /&gt;&lt;BR /&gt;Hope this helps&lt;BR /&gt;Chris</description>
      <pubDate>Tue, 23 Oct 2001 09:19:51 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599463#M716014</guid>
      <dc:creator>Christopher McCray_1</dc:creator>
      <dc:date>2001-10-23T09:19:51Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599464#M716015</link>
      <description>Thanks Chris&lt;BR /&gt;I didn't mention that the VG-Lockdisks reside on the XP in the second Datacenter, the ones the primary nodes can't reach anymore.&lt;BR /&gt;I'll try to attach the config once more.</description>
      <pubDate>Tue, 23 Oct 2001 13:02:08 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599464#M716015</guid>
      <dc:creator>Paul Barmettler</dc:creator>
      <dc:date>2001-10-23T13:02:08Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599465#M716016</link>
      <description>I want to make sure I understand the physical layout.  Is it one of the following:&lt;BR /&gt;&lt;BR /&gt;1) the servers (alt and pri) in one datacenter and the xp in the other?&lt;BR /&gt;2) the primary servers in one data center and the alts in the other;  in this case which servers is the xp co-located with?&lt;BR /&gt;&lt;BR /&gt;Sorry, but I must have missed the xp location part, but it would help a lot if you can answer the above.  Thanks.&lt;BR /&gt;&lt;BR /&gt;Chris</description>
      <pubDate>Tue, 23 Oct 2001 13:15:17 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599465#M716016</guid>
      <dc:creator>Christopher McCray_1</dc:creator>
      <dc:date>2001-10-23T13:15:17Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599466#M716017</link>
      <description>You should really provide the syslog of both nodes to give us a better picture of the event. &lt;BR /&gt;&lt;BR /&gt;One thing what I believe what you might oversee is that both nodes need to get access to BOTH cluster lock disks. Under specific circumstances (ie. the return code of the system call to access the cluster lock indicates an I/O error or powerfailure of the disk) SG requires only one of the two lock disk to form a cluster.&lt;BR /&gt;&lt;BR /&gt;Without seeing the syslogs, I dare to maintain that the primary node got the cluster lock (of the alternate data center?) and the alternate did not and therefore performed a TOC. The syslogs will give us the details. &lt;BR /&gt;&lt;BR /&gt;Carsten</description>
      <pubDate>Tue, 23 Oct 2001 14:45:02 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599466#M716017</guid>
      <dc:creator>Carsten Krege</dc:creator>
      <dc:date>2001-10-23T14:45:02Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599467#M716018</link>
      <description>That's what I was thinking and hope I was conveying, Karsten.  I was thinking that the primary servers some how had connectivity to a cluster lock disk.  Thanks for clearing that up for me; I fear I was too vague.&lt;BR /&gt;&lt;BR /&gt;It is possible that this is the case, Neuhaus, but please send us the logs Karsten mentioned and thanks.&lt;BR /&gt;&lt;BR /&gt;Chris&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Tue, 23 Oct 2001 14:51:04 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599467#M716018</guid>
      <dc:creator>Christopher McCray_1</dc:creator>
      <dc:date>2001-10-23T14:51:04Z</dc:date>
    </item>
    <item>
      <title>Re: Desaster Test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599468#M716019</link>
      <description>Thanks for your responses.&lt;BR /&gt;I guess Carsten you are right. One of my colleagues told me, that he has seen the message on the console of an alternative node, tha he was not able to optain the cluster lock disk. Pobbably because there was too much time between cutting the LAN cables and cutting the FC cables.&lt;BR /&gt;Sorry I cant' provide syslogs, because I din't save them before the next reboot!!&lt;BR /&gt;For your understanding I attach the physical layout.</description>
      <pubDate>Wed, 24 Oct 2001 10:53:18 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/desaster-test/m-p/2599468#M716019</guid>
      <dc:creator>Paul Barmettler</dc:creator>
      <dc:date>2001-10-24T10:53:18Z</dc:date>
    </item>
  </channel>
</rss>

