<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Cluster member crash has high impact on other nodes in the cluster in Operating System - OpenVMS</title>
    <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684214#M100065</link>
    <description>do also consider re-configuring to put the dump file on local storage (SAN).&lt;BR /&gt;&lt;BR /&gt;don't rush to make changes.</description>
    <pubDate>Wed, 08 Sep 2010 08:37:48 GMT</pubDate>
    <dc:creator>Ian Miller.</dc:creator>
    <dc:date>2010-09-08T08:37:48Z</dc:date>
    <item>
      <title>Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684204#M100055</link>
      <description>Hello,&lt;BR /&gt;&lt;BR /&gt;We have a VMS cluster of 6 x I64 servers and 2 x Alpha servers. We use ethernet/LAN as Cluster interconnect. &lt;BR /&gt;One I64 server crashed today but it had a hudge impact on the other members in the cluster. Many processes on the other nodes came in RWSCS state.&lt;BR /&gt;We use HBMM.&lt;BR /&gt;&lt;BR /&gt;Is this normal?&lt;BR /&gt;Which sysgen parameter should I change to avoid this or minimize this behaviour.&lt;BR /&gt;&lt;BR /&gt;Also I got strange message on the I64 Console of the failing node.&lt;BR /&gt;&lt;BR /&gt;**** Unable to write header, dump will probably be unusable ****PGQBT-E-Transport Error IO[11]: STS 0x2, SCSISTS 0x0, STSFLG 0x0, STATEFLG 0x0&lt;BR /&gt;&lt;BR /&gt;Does anyone know what this mean?&lt;BR /&gt;&lt;BR /&gt;Toine</description>
      <pubDate>Tue, 07 Sep 2010 14:50:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684204#M100055</guid>
      <dc:creator>Toine_1</dc:creator>
      <dc:date>2010-09-07T14:50:44Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684205#M100056</link>
      <description>You do have a SAN here (based on that "PGQBT"), right?&lt;BR /&gt;&lt;BR /&gt;Without the SAN, there is a reasonable chance that the (presumably) gigabit ethernet LAN is overloaded; eight hosts and some number of HBVS full-merge operations is a whole lot of network traffic, after all.  &lt;BR /&gt;&lt;BR /&gt;Even with the SAN, you might have a plugged network.&lt;BR /&gt;&lt;BR /&gt;Based on the PGQBT boot driver diagnostic, there was apparently some sort of a SAN error here during the crash.  The box apparently couldn't get to the SAN or to the storage controller or to the disk.&lt;BR /&gt;&lt;BR /&gt;I'd dispense with the tuning effort at least temporarily, and start investigating the steady-state and the HBVS recovery network loading (with T4 as well as with a network monitor hanging off a "mirror" port on your network switch), with an investigation of what hardware is present here, and I'd look at adding links and faster interconnects. &lt;BR /&gt;&lt;BR /&gt;Definitely check for ECO kits.&lt;BR /&gt;&lt;BR /&gt;And check the boot device.&lt;BR /&gt;&lt;BR /&gt;And check the error logs.&lt;BR /&gt;&lt;BR /&gt;And if you have support, call HP.</description>
      <pubDate>Tue, 07 Sep 2010 15:11:53 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684205#M100056</guid>
      <dc:creator>Hoff</dc:creator>
      <dc:date>2010-09-07T15:11:53Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684206#M100057</link>
      <description>What is the storage and networking configuration?  Are there multiple network adaptors?  Are speed and duplex configured propertly for all paths?  What is the network speed?  Is there a SAN and does each system have HBAs?&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Tue, 07 Sep 2010 15:37:27 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684206#M100057</guid>
      <dc:creator>Andy Bustamante</dc:creator>
      <dc:date>2010-09-07T15:37:27Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684207#M100058</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;I have a SAN of two EVA4400 and two HBA cards in each server. So each server has 4 paths to each disk.&lt;BR /&gt;The EVA boxes are located in two computerroom.&lt;BR /&gt;&lt;BR /&gt;Two brocade switches in each computerroom for each EVA4400.&lt;BR /&gt;&lt;BR /&gt;Each server has two Gigabit LAN interfaces.&lt;BR /&gt;One Gigabit interface is connected to a dedicated switch for the cluster communication.&lt;BR /&gt;&lt;BR /&gt;The error count on the PE device was increased during this problem.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;I use Host based mini merge but can it be that during this mini merge some I/O are blocked.&lt;BR /&gt;It was also strange that all processes that are using sockets connections were in RWSCS state.&lt;BR /&gt;Also via Telnet no one could logon to the remaining members for a short period.&lt;BR /&gt;&lt;BR /&gt;$ show shadow sys$sysdevice&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;_DSA100:  Volume Label: I64VMS&lt;BR /&gt;  Virtual Unit State:   Steady State&lt;BR /&gt;  Enhanced Shadowing Features in use:&lt;BR /&gt;        Host-Based Minimerge (HBMM)&lt;BR /&gt;&lt;BR /&gt;  VU Timeout Value  16777215    VU Site Value          1&lt;BR /&gt;  Copy/Merge Priority   5000    Mini Merge       Enabled&lt;BR /&gt;  Recovery Delay Per Served Member                    30&lt;BR /&gt;  Merge Delay Factor     200    Delay Threshold      200&lt;BR /&gt;&lt;BR /&gt;  HBMM Policy&lt;BR /&gt;    HBMM Reset Threshold: 6000000&lt;BR /&gt;    HBMM Master lists:&lt;BR /&gt;      Up to any 3 of the nodes: NVR,NVC,NVE,NVJ Multiuse: 0&lt;BR /&gt;    HBMM bitmaps are active on NVJ,NVC,NVE&lt;BR /&gt;  HBMM Reset Count      49       Last Reset     7-SEP-2010 12:23:49.90&lt;BR /&gt;    Modified blocks since last bitmap reset: 5976239&lt;BR /&gt;&lt;BR /&gt;  Device $1$DGA100              Master Member&lt;BR /&gt;    Read Cost              2    Site 1&lt;BR /&gt;    Member Timeout       120&lt;BR /&gt;&lt;BR /&gt;  Device $1$DGA200&lt;BR /&gt;    Read Cost             42    Site 2&lt;BR /&gt;    Member Timeout       120&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Toine</description>
      <pubDate>Tue, 07 Sep 2010 16:16:19 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684207#M100058</guid>
      <dc:creator>Toine_1</dc:creator>
      <dc:date>2010-09-07T16:16:19Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684208#M100059</link>
      <description>RWSCS is a cluster wait.  &lt;BR /&gt;&lt;BR /&gt;Short RWSCS waits are entirely normal.&lt;BR /&gt;&lt;BR /&gt;Longer RWSCS can indicate a blocked network.  Or blocked locking.  Or cluster credit exhaustion.  Or lock manager flailing.  I'd also expect that a stuffed-up SAN could also trigger this resource wait state, too.  &lt;BR /&gt;&lt;BR /&gt;And that PGQBT SAN error is worth investigation.&lt;BR /&gt;&lt;BR /&gt;You're going to have to instrument the cluster and the LAN, via wireshark and T4, or analogous.&lt;BR /&gt;&lt;BR /&gt;You're also going to have to investigate the error logs.&lt;BR /&gt;&lt;BR /&gt;Also the power stability, and the contents of the network and storage server logs.&lt;BR /&gt;&lt;BR /&gt;If you have HP support available, start down that path now.  (Your management paid good money for that, too.)</description>
      <pubDate>Tue, 07 Sep 2010 16:29:38 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684208#M100059</guid>
      <dc:creator>Hoff</dc:creator>
      <dc:date>2010-09-07T16:29:38Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684209#M100060</link>
      <description>Thank you Hoff,&lt;BR /&gt;&lt;BR /&gt;What I also saw that there was during 6 minutes a queue length of 20 on the system disk of the I64 servers.&lt;BR /&gt;We use one system disk for all I64 servers.&lt;BR /&gt;&lt;BR /&gt;I have logged a call at HP and I hope we will find teh root cause.&lt;BR /&gt;&lt;BR /&gt;Regards,&lt;BR /&gt;&lt;BR /&gt;Toine,</description>
      <pubDate>Tue, 07 Sep 2010 16:37:21 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684209#M100060</guid>
      <dc:creator>Toine_1</dc:creator>
      <dc:date>2010-09-07T16:37:21Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684210#M100061</link>
      <description>Toine,&lt;BR /&gt;&lt;BR /&gt;  Please post the output of:&lt;BR /&gt;&lt;BR /&gt;$ MCR SYSGEN SHOW/CLUSTER&lt;BR /&gt;&lt;BR /&gt;The tradeoff with failures in a cluster is around how long to wait before deciding an apparently lost node is really lost. One key parameter is RECNXINTERVAL. If another node stops responding, surviving nodes wait that many seconds to see if it reappears.&lt;BR /&gt;&lt;BR /&gt;  If RECNXINTERVAL is too high, the whole cluster can freeze for that long before even attempting to reform without the missing node. If the value is too low, some transient event on your cluster interconnect can cause the cluster to kick a node out unnecessarily.&lt;BR /&gt;&lt;BR /&gt;  Another issue which affects the timing of recovery from failure is where your locks are mastered. This is mostly controlled by LOCKDIRWT. Much of the time of a cluster transition is working out which lock resources have been "lost" (because they were mastered on the lost node), deciding which node will take over that resource, and reconciling the states of any interested locks on surviving cluster nodes.&lt;BR /&gt;&lt;BR /&gt;  If you happened to have a large lock tree, with lots of intercluster activity mastered on the node which crashed, then it could take substantial time (order minutes) to reconstruct the tree. Processes waiting on locks against lost resources will wait in RWSCS state while the states are sorted out. There's not a lot you can do about his, except perhaps to make sure, if you have multiple, large lock trees, that they are not concentrated on a single node.&lt;BR /&gt;&lt;BR /&gt;  Find out what lock trees normally live on your system, and how they are distributed. If they're all on one node, and that node is lost, you have to rebuild them all.</description>
      <pubDate>Tue, 07 Sep 2010 20:52:04 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684210#M100061</guid>
      <dc:creator>John Gillings</dc:creator>
      <dc:date>2010-09-07T20:52:04Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684211#M100062</link>
      <description>Hello John,&lt;BR /&gt;&lt;BR /&gt;You are correct perhaps the RECNXINTERVAL is too high 60 seconds in my cluster.&lt;BR /&gt;&lt;BR /&gt;I must also tell you that the two Integrity servers with the highest LOCKDIRWT didn't crash. But I will check the lock remastering.&lt;BR /&gt;&lt;BR /&gt;Below the sysgen parameters&lt;BR /&gt;&lt;BR /&gt;$ mc sysgen sho/cluster&lt;BR /&gt;&lt;BR /&gt;Parameters in use: Active&lt;BR /&gt;Parameter Name            Current    Default     Min.       Max.   Unit  Dynamic&lt;BR /&gt;--------------            -------    -------   -------    -------  ----  -------&lt;BR /&gt;VAXCLUSTER                      2          1         0          2 Coded-valu&lt;BR /&gt;EXPECTED_VOTES                 10          1         1        127 Votes&lt;BR /&gt;VOTES                           2          1         0        127 Votes&lt;BR /&gt;DISK_QUORUM     "                "    "    "    "    "     "ZZZZ" Ascii&lt;BR /&gt;QDSKVOTES                       1          1         0        127 Votes&lt;BR /&gt;QDSKINTERVAL                    3          3         1      32767 Seconds&lt;BR /&gt;ALLOCLASS                       1          0         0        255 Pure-numbe&lt;BR /&gt;LOCKDIRWT                       6          0         0        255 Pure-numbe&lt;BR /&gt;CLUSTER_CREDITS               128         32        10        128 Credits&lt;BR /&gt;NISCS_CONV_BOOT                 0          0         0          1 Boolean&lt;BR /&gt;NISCS_LOAD_PEA0                 1          0         0          1 Boolean&lt;BR /&gt;NISCS_USE_LAN                   1          1         0          1 Boolean&lt;BR /&gt;NISCS_USE_UDP                   0          0         0          1 Boolean&lt;BR /&gt;MSCP_LOAD                       1          0         0      16384 Coded-valu&lt;BR /&gt;TMSCP_LOAD                      0          0         0          3 Coded-valu&lt;BR /&gt;MSCP_SERVE_ALL                  1          4         0         -1 Bit-Encode&lt;BR /&gt;TMSCP_SERVE_ALL                 0          0         0         -1 Bit-Encode&lt;BR /&gt;MSCP_BUFFER                 16384       1024       256         -1 Coded-valu&lt;BR /&gt;MSCP_CREDITS                  128         32         2       1024 Coded-valu&lt;BR /&gt;TAPE_ALLOCLASS                  0          0         0        255 Pure-numbe&lt;BR /&gt;NISCS_MAX_PKTSZ              8192       8192       576       9180 Bytes&lt;BR /&gt;CWCREPRC_ENABLE                 1          1         0          1 Bitmask    D&lt;BR /&gt;RECNXINTERVAL                  60         20         1      32767 Seconds    D&lt;BR /&gt;NISCS_PORT_SERV                 0          0         0        256 Bitmask    D&lt;BR /&gt;NISCS_UDP_PORT                  0          0         0      65535 Pure-numbe D&lt;BR /&gt;MSCP_CMD_TMO                    0          0         0 2147483647 Seconds    D&lt;BR /&gt;LOCKRMWT                        5          5         0         10 Pure-numbe&lt;BR /&gt;&lt;BR /&gt;Toine</description>
      <pubDate>Tue, 07 Sep 2010 21:05:59 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684211#M100062</guid>
      <dc:creator>Toine_1</dc:creator>
      <dc:date>2010-09-07T21:05:59Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684212#M100063</link>
      <description>Toine,&lt;BR /&gt;&lt;BR /&gt;&amp;gt;perhaps the RECNXINTERVAL is too high 60 seconds &lt;BR /&gt;&lt;BR /&gt;Don't assume that! Someone has set RECNXINTERVAL up from default, hopefully with good reason.&lt;BR /&gt;&lt;BR /&gt;That means if a node loses power, is disconnected, or crashes, you will expierience a cluster state transition of at least 60 seconds. BUT depending on your network infrastructure, and business needs, that may be perfectly reasonable.&lt;BR /&gt;&lt;BR /&gt;Consider, if cluster nodes are separated by a long distance, the cluster interconnect may go through various network boxes. If the reboot time for one of those boxes is (say) 30 seconds, you may WANT a relatively high RECNXINTERVAL so your cluster will survive an expected network outage. &lt;BR /&gt;&lt;BR /&gt;As long as the business can tolerate up to a 60 second pause if there's a network transient, that may be preferable to having nodes kicked out unnecessarily.&lt;BR /&gt;&lt;BR /&gt;Only you, and your internal business customers can decide the best tradeoff for your systems. &lt;BR /&gt;&lt;BR /&gt;If you need shorter transitions, one way to allow you to reduce RECNXINTERVAL is to have multiple cluster interconnect paths. That way, even if you lose connectivity on one path, the remaining one(s) will keep the cluster together. Often modern systems have several network adapters, some of which may be unused. Perhaps you can connect all nodes using "spare" adapters through a private switch. Watch out for single points of failure.&lt;BR /&gt;&lt;BR /&gt;As always, you need to balance costs and business needs.</description>
      <pubDate>Tue, 07 Sep 2010 23:36:57 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684212#M100063</guid>
      <dc:creator>John Gillings</dc:creator>
      <dc:date>2010-09-07T23:36:57Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684213#M100064</link>
      <description>Toine&lt;BR /&gt;&lt;BR /&gt;You said&lt;BR /&gt;"But I will check the lock remastering."&lt;BR /&gt;&lt;BR /&gt;Be careful with the SDA extension LCK&lt;BR /&gt;&lt;BR /&gt;Usually, we assume that we can do nearly whatever we want in SDA with no harm. Just looking at memory locations is innocent.&lt;BR /&gt;&lt;BR /&gt;This is not correct, as a &lt;BR /&gt;SDA&amp;gt; lck remaster...&lt;BR /&gt;can use a lot of CPU, generate processes in RWSCS, RWCLU...&lt;BR /&gt;&lt;BR /&gt;Of course, &lt;BR /&gt;SDA&amp;gt; lck stat/toptrees=10&lt;BR /&gt;is "innocent"&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Wed, 08 Sep 2010 06:01:52 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684213#M100064</guid>
      <dc:creator>labadie_1</dc:creator>
      <dc:date>2010-09-08T06:01:52Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684214#M100065</link>
      <description>do also consider re-configuring to put the dump file on local storage (SAN).&lt;BR /&gt;&lt;BR /&gt;don't rush to make changes.</description>
      <pubDate>Wed, 08 Sep 2010 08:37:48 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684214#M100065</guid>
      <dc:creator>Ian Miller.</dc:creator>
      <dc:date>2010-09-08T08:37:48Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684215#M100066</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;Thank you all for the good answers. very helpfull as always.&lt;BR /&gt;&lt;BR /&gt;HP advised us to disable the CPE monitoring via a sysgen paramater as a work around.&lt;BR /&gt; &lt;BR /&gt;&lt;BR /&gt;SYSGEN&amp;gt; use active&lt;BR /&gt;SYSGEN&amp;gt; set crd_control %x80016&lt;BR /&gt;SYSGEN&amp;gt; write active  &lt;BR /&gt;SYSGEN&amp;gt; use current&lt;BR /&gt;SYSGEN&amp;gt; set crd_control %x80016&lt;BR /&gt;SYSGEN&amp;gt; show crd_control&lt;BR /&gt;SYSGEN&amp;gt; write current&lt;BR /&gt;SYSGEN&amp;gt; exit&lt;BR /&gt;&lt;BR /&gt;Regards,&lt;BR /&gt;&lt;BR /&gt;Toine</description>
      <pubDate>Wed, 08 Sep 2010 13:30:33 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684215#M100066</guid>
      <dc:creator>Toine_1</dc:creator>
      <dc:date>2010-09-08T13:30:33Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684216#M100067</link>
      <description>The suggested remediation from HP implies that there are buckets of memory errors arising here, and that's something I've definitely encountered on a few Integrity and AlphaServer boxes over the years.  RAS features or not, memory errors can cause instabilities on both Integrity and AlphaServer boxes.  On any box, for that matter.  And the errors don't always get overtly logged; you have to go look for them.</description>
      <pubDate>Wed, 08 Sep 2010 14:43:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684216#M100067</guid>
      <dc:creator>Hoff</dc:creator>
      <dc:date>2010-09-08T14:43:44Z</dc:date>
    </item>
    <item>
      <title>Re: Cluster member crash has high impact on other nodes in the cluster</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684217#M100068</link>
      <description>Toine,&lt;BR /&gt;  &lt;BR /&gt;&amp;gt; SYSGEN&amp;gt; set crd_control %x80016&lt;BR /&gt;&lt;BR /&gt;  Hmmm, someone in HP support needs to be taught "Balmer's Rule" (someone, somewhere on the planet may laugh... :-)&lt;BR /&gt;&lt;BR /&gt;You've adjusted the parameter value correctly (USE/SET/WRITE), but if you don't want a surprise sometime in the future when you've forgotten about this thread, you should also add that SET command to MODPARAMS.DAT, commented, with your name, date, reference to the HP service case, and maybe even the URL of this thread.&lt;BR /&gt;&lt;BR /&gt;System configurations are complex things, and changes often have unintended consequences. As an example, in this case it may have been helpful to know why your cluster has a non-default RECNXINTERVAL. Who decided one the value, when and why? If your system doesn't have a clearly documented MODPARAMS.DAT, please start now.&lt;BR /&gt;&lt;BR /&gt;Also, it just occurred to me, that since your cluster has 8 nodes, you should check that RECNXINTERVAL is consistent across all nodes. I'd expect that the resultant delay would be the longest across the cluster (of course, inconsistencies like that SHOULD be detected and notified at cluster formation, but engineering has never considered it important enough to expend resources implementing a proper cluster consistency check :-(</description>
      <pubDate>Wed, 08 Sep 2010 20:38:13 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/cluster-member-crash-has-high-impact-on-other-nodes-in-the/m-p/4684217#M100068</guid>
      <dc:creator>John Gillings</dc:creator>
      <dc:date>2010-09-08T20:38:13Z</dc:date>
    </item>
  </channel>
</rss>

