<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Quorum Disk Failure in Operating System - OpenVMS</title>
    <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184440#M61896</link>
    <description>Sorry, two irritating typos:&lt;BR /&gt;&lt;BR /&gt;"@ nodes"  should read  "2 nodes" &lt;BR /&gt;" +&amp;gt; "        "    "     " =&amp;gt; "&lt;BR /&gt;&lt;BR /&gt;... I sometimes (have to) work on systems with different keyboard layouts.  It should have been forbidden, but then, who should be allowed to declare "THE" correct layout?</description>
    <pubDate>Sat, 07 Feb 2004 08:36:42 GMT</pubDate>
    <dc:creator>Jan van den Ende</dc:creator>
    <dc:date>2004-02-07T08:36:42Z</dc:date>
    <item>
      <title>Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184435#M61891</link>
      <description>We have a two node cluster with a quorum disk and are familiar with what happens when a node is shut down or fails - the other node goes into a cluster reconfigure along with issuing Operator msgs.&lt;BR /&gt;&lt;BR /&gt;What happens if the Quorum disk fails?&lt;BR /&gt;&lt;BR /&gt;Is the scenario the same on the hosts - cluster reconfigure with msgs referencing the loss of the quorum disk or does something different happen?&lt;BR /&gt;&lt;BR /&gt;What happens when the failed Quorum disk is made available again?&lt;BR /&gt;&lt;BR /&gt;Do the hosts automatically recongize the appearance of the quorum disk and use it (with appropriate msgs)?&lt;BR /&gt;&lt;BR /&gt;In other words, how does the loss/restore of a quorum disk compare to loss/restore of a host node?&lt;BR /&gt;&lt;BR /&gt;One more question (for extra points!).  What happens when the quorum disk "temporarily" disappears?  We have our quorum disk on an HP SW SAN.  If we make a Zoning or Presentation change that affects the quorum disk, besides VMS going through a Mount Verify, what does the host recovery look like?&lt;BR /&gt;&lt;BR /&gt;Thanks much</description>
      <pubDate>Thu, 05 Feb 2004 13:16:46 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184435#M61891</guid>
      <dc:creator>Jack Trachtman</dc:creator>
      <dc:date>2004-02-05T13:16:46Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184436#M61892</link>
      <description>the quorum disk watcher sees all! :-)&lt;BR /&gt;(in QDISKINTERVAL seconds)&lt;BR /&gt;You will see messages about the quorum disk being unavailable and as long as there are enough votes then the cluster will continue. When  the qdsk returns it will be recognised as the quorum disk by the presence of the QUORUM.DAT and the votes (QDSKVOTES) counted. &lt;BR /&gt;&lt;BR /&gt;I assume both systems are directly connected to the quorum disk.</description>
      <pubDate>Thu, 05 Feb 2004 13:25:16 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184436#M61892</guid>
      <dc:creator>Ian Miller.</dc:creator>
      <dc:date>2004-02-05T13:25:16Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184437#M61893</link>
      <description>Well, except for its obvious passiveness, the quorumdisk should be considered as 'just' another node, although, due to the mentioned passivenees, much slower to respond. &lt;BR /&gt;It is indirect. The first node to reconnect checks the quorumdisk for recent 'stamps' by the other node(s). They are not there, so it simply leaves its own imprint. The second one finds teh stamp (and it better be of a known node). It also leaves its trace, and the next time #1 comes along it can conclude that the quorum disk is a valid member again.&lt;BR /&gt;&lt;BR /&gt;Even IF (unwanted situation) the departure of the quorumdisk leads to a loss of quorum (the cluster 'hangs') this mechanism is above the hang, and if the return of Qdsk suffices to regain quorum, that WILL be recognised, and the hang will be over.&lt;BR /&gt;&lt;BR /&gt;jan</description>
      <pubDate>Fri, 06 Feb 2004 14:15:05 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184437#M61893</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-02-06T14:15:05Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184438#M61894</link>
      <description>In a 2 node cluster, one can avoid a hang if the quorum disk fails by setting Votes=2 on each node, and qdskvotes=1 so expected_votes=5. This way, if either node goes down, the requisite minimum of 3 votes is held and if the quorum disk goes bye-bye, 4 votes are held.&lt;BR /&gt;Only in the event of a double failure does the surviving node hang.&lt;BR /&gt;Maybe that will help someone. &lt;BR /&gt;john</description>
      <pubDate>Fri, 06 Feb 2004 16:33:27 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184438#M61894</guid>
      <dc:creator>John Eerenberg</dc:creator>
      <dc:date>2004-02-06T16:33:27Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184439#M61895</link>
      <description>John,&lt;BR /&gt;&lt;BR /&gt;works fine, but not even needed.&lt;BR /&gt;@ nodes each 1 vote + qdsk 1 vote = 3 votes expected. Any single voter gone leaves 2 vote +&amp;gt; quorum maintained.&lt;BR /&gt;&lt;BR /&gt;The abovementioned temporary hang (and resume) would occur at one node out (eg, maintenace) and THEN having your SAN disconnecting and reconnecting the qdsk.&lt;BR /&gt;&lt;BR /&gt;hth&lt;BR /&gt;&lt;BR /&gt;Jan</description>
      <pubDate>Sat, 07 Feb 2004 05:14:28 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184439#M61895</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-02-07T05:14:28Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184440#M61896</link>
      <description>Sorry, two irritating typos:&lt;BR /&gt;&lt;BR /&gt;"@ nodes"  should read  "2 nodes" &lt;BR /&gt;" +&amp;gt; "        "    "     " =&amp;gt; "&lt;BR /&gt;&lt;BR /&gt;... I sometimes (have to) work on systems with different keyboard layouts.  It should have been forbidden, but then, who should be allowed to declare "THE" correct layout?</description>
      <pubDate>Sat, 07 Feb 2004 08:36:42 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184440#M61896</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-02-07T08:36:42Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184441#M61897</link>
      <description>Hi ,&lt;BR /&gt;&lt;BR /&gt;I agree with Jan. Why setting votes=2 for cluster members ? It works fine for me with votes=1 .&lt;BR /&gt;&lt;BR /&gt;Best regards,&lt;BR /&gt;Lokesh</description>
      <pubDate>Mon, 09 Feb 2004 06:29:29 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184441#M61897</guid>
      <dc:creator>Lokesh_2</dc:creator>
      <dc:date>2004-02-09T06:29:29Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184442#M61898</link>
      <description>If you give 2 votes to each cluster member and 1 to the quorum disk, you have the advantage that&lt;BR /&gt;IF&lt;BR /&gt;1) node 1 is stopped with REMOVE_NODE&lt;BR /&gt;2) the quorum disk gets lost after 1) has completed&lt;BR /&gt;THEN&lt;BR /&gt;your cluster is still alive (because a minority of 1 vote left the cluster with total votes equal to 3, so 2 votes left).&lt;BR /&gt;&lt;BR /&gt;If you have 1-1-1, the cluster would hang until the disk is replaced or the second node is rebooted.</description>
      <pubDate>Mon, 09 Feb 2004 07:01:15 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184442#M61898</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-02-09T07:01:15Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184443#M61899</link>
      <description>Touchee Wim!&lt;BR /&gt;&lt;BR /&gt;Indeed, that IS the reason to use 2-2-1.&lt;BR /&gt;&lt;BR /&gt;Then the issue of SAN disconnecting/reconnecting disappears nearly completely: Only if one node left WITHOUT adjustment (eigther by crash or by operator forgetting "remove_node"), THEN if the SAN connection disappears BEFORE a SET CLUSTER/EXPECTED,only THEN will the hang still occur. Should be very rare.&lt;BR /&gt;Jan</description>
      <pubDate>Mon, 09 Feb 2004 08:10:45 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184443#M61899</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-02-09T08:10:45Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184444#M61900</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;Thanks for explaning the advantage of 2-2-1 . I will note it down.&lt;BR /&gt;&lt;BR /&gt;Best regards,&lt;BR /&gt;Lokesh</description>
      <pubDate>Mon, 09 Feb 2004 09:25:06 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184444#M61900</guid>
      <dc:creator>Lokesh_2</dc:creator>
      <dc:date>2004-02-09T09:25:06Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184445#M61901</link>
      <description>Just some icing on the cake . . .&lt;BR /&gt;a parallel to the Remove_Node option is when one has a system crash. &lt;BR /&gt;I set up the 2-2-1 votes so that if  Node A crashes, then I can go to Node B's console and adjust quorum:&lt;BR /&gt;^p&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; dep sirr c&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; cont&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; q&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; ^z&lt;BR /&gt;&lt;BR /&gt;Now I can lose the quorum disk too and the surviving node can stay up as a standalone system.&lt;BR /&gt;&lt;BR /&gt;BE very CAREFUL when using the above console commands. If I entered the above incorrectly *and/or* one makes a typo, the surviving node can crash as well. Test it before needing it.&lt;BR /&gt;&lt;BR /&gt;john</description>
      <pubDate>Mon, 09 Feb 2004 13:54:07 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184445#M61901</guid>
      <dc:creator>John Eerenberg</dc:creator>
      <dc:date>2004-02-09T13:54:07Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184446#M61902</link>
      <description>John,&lt;BR /&gt;&lt;BR /&gt;just THAT would be your emergency escape if you get to HANG, eg 1-1-1 and losing 2.&lt;BR /&gt;In the 2-2-1 scheme, if one crashes, you DON'T have to IPC the other node. If your cluster is NOT YET hung, but you fear the crashed node might be down for whatever you define as 'a prolongued time', then you get to exactly the same result by SET CLUSTER/EXPECTED from any sufficiently privileged (CMKRNL, SYSNAM, &amp;amp; SYSLCK) process.&lt;BR /&gt;Alternatively, AMDS or Availability Manager if installed and configured correctly will do the same for you, and also in a much more controlled way.&lt;BR /&gt;&lt;BR /&gt;Then again, I ALWAYS carry a note with the IPC sequence in my wallet. In 20 years I needed it twice, and then carrying a small note is invaluable!&lt;BR /&gt;&lt;BR /&gt;Jan</description>
      <pubDate>Mon, 09 Feb 2004 14:38:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184446#M61902</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-02-09T14:38:14Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184447#M61903</link>
      <description>Agreed!&lt;BR /&gt;I guess I still have an old school mentality since I still remember having to hit bit 28 to boot TOPS10 on a KI10 (are there any KI10's left on the planet??? or TOPS10 for that matter?)&lt;BR /&gt;&lt;BR /&gt;&amp;gt; Then again, I ALWAYS carry a note with the IPC sequence in my wallet. In 20 years I needed it twice, and then carrying a small note is invaluable!&lt;BR /&gt;&lt;BR /&gt;Once was all I really needed it so you have me beat. :-)&lt;BR /&gt;Kept it on paper as well until I recently bought an iPaq 1945; something about paper was more reassuring though . . .&lt;BR /&gt;</description>
      <pubDate>Tue, 10 Feb 2004 10:25:32 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184447#M61903</guid>
      <dc:creator>John Eerenberg</dc:creator>
      <dc:date>2004-02-10T10:25:32Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184448#M61904</link>
      <description>Two questions:&lt;BR /&gt;&lt;BR /&gt;1) This is probably obvious, but what&lt;BR /&gt;does IPC stand for in the previous msgs&lt;BR /&gt;context?&lt;BR /&gt;&lt;BR /&gt;2) In John's msg, he has "&amp;gt;&amp;gt;&amp;gt; dep sirr c",&lt;BR /&gt;which is a deposit cmd, but what is&lt;BR /&gt;"sirr" and why the value "c"?&lt;BR /&gt;&lt;BR /&gt;Thanks</description>
      <pubDate>Tue, 10 Feb 2004 11:23:00 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184448#M61904</guid>
      <dc:creator>Jack Trachtman</dc:creator>
      <dc:date>2004-02-10T11:23:00Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184449#M61905</link>
      <description>sirr = software interrupt request register.&lt;BR /&gt;C = ipl at which to request the interrupt i.e. 12&lt;BR /&gt;The handler for software interrupts at 12 is the recalculate quorum routine which has a prompt of IPC (interrupt prio C). &lt;BR /&gt;&lt;BR /&gt;The use of this routine is not recommended nowadays. Use AMDS or Availability Manager. If you have not got AMDS/AvailMgr setup do so now!&lt;BR /&gt; They are invaluable for many reasons.&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Tue, 10 Feb 2004 12:42:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184449#M61905</guid>
      <dc:creator>Ian Miller.</dc:creator>
      <dc:date>2004-02-10T12:42:01Z</dc:date>
    </item>
    <item>
      <title>Re: Quorum Disk Failure</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184450#M61906</link>
      <description>All,&lt;BR /&gt;&lt;BR /&gt;Remember when you fix the quorum by hand, you have the gate open for a split cluster.&lt;BR /&gt;&lt;BR /&gt;E.g. an interbuilding cluster with a quorum station in 1 building. The quorum station + 1 node in the same building go down. The remaining node in the other building is alive but blocked because of missing quorum. You adjust the quorum and the single node continues. The interbuilding link is down and the nodes in the other building resume activity (e.g. after power failure). You now have 2 clusters.&lt;BR /&gt;&lt;BR /&gt;That's why my procedures refuse to boot if I don't see all disks in both buildings.</description>
      <pubDate>Wed, 11 Feb 2004 02:32:12 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/quorum-disk-failure/m-p/3184450#M61906</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-02-11T02:32:12Z</dc:date>
    </item>
  </channel>
</rss>

