<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: fail over test in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264731#M569296</link>
    <description>Ross,&lt;BR /&gt;&lt;BR /&gt;I am having a 11.0 server in the exact same configuration you have and it is running pretty smooth without any problems.It is one of our legacy applications which has not been migrated to 11i. &lt;BR /&gt;We ahve however had numerous problems with the VA and have stopped using it in favour of a XP1024. &lt;BR /&gt;Let me poke around my server and see if I can come up with something you can use.</description>
    <pubDate>Fri, 30 Apr 2004 17:30:19 GMT</pubDate>
    <dc:creator>hari jayaram_1</dc:creator>
    <dc:date>2004-04-30T17:30:19Z</dc:date>
    <item>
      <title>fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264726#M569291</link>
      <description>We did a fail over test of a N4000 server running informix dynamic server ver. 7.3 connected by two fc wires to two brocade switches with redundent paths going to a va7400 san box. &lt;BR /&gt;We pulled out one controller from the san box and crashed the database.&lt;BR /&gt;We then pulled out the other san controller and crashed the database again.&lt;BR /&gt;Then we turned off one of the brocade switches and crashed the database, after that we did the same thing to the other brocade switch and once &lt;BR /&gt;again the database came down.&lt;BR /&gt;Doing a pvdisplay -v /dev/vg05 and vg06 shows that alternate paths exist so does anyone know why informix, running on HPUX 11.0 would not fail over properly?</description>
      <pubDate>Fri, 30 Apr 2004 16:52:04 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264726#M569291</guid>
      <dc:creator>ROSS HANSON</dc:creator>
      <dc:date>2004-04-30T16:52:04Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264727#M569292</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;Could you please advice if you are seeing any error messages.</description>
      <pubDate>Fri, 30 Apr 2004 17:01:53 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264727#M569292</guid>
      <dc:creator>hari jayaram_1</dc:creator>
      <dc:date>2004-04-30T17:01:53Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264728#M569293</link>
      <description>Hi Ross,&lt;BR /&gt;&lt;BR /&gt;You're under the assumption that a disk I/O failure should cause a failover. Not gonna happen! MC/SG needs to shut the SW down when it wants to failover &amp;amp; how is it gonna do that if it's I/O is gone? The same kind of rule applies to the CPUs - BUT in that case the box is gonna panic &amp;amp; TOC &amp;amp; THEN the other node will pick up the ball &amp;amp; start up the package. Ditto with power supplies.&lt;BR /&gt;Bottom line is that MC/SG is *not* a continously available solution - its *highly* available. For continous you need to spend *far* more $ with HP. They can do it, but not for the price of MC/SG. &lt;BR /&gt;&lt;BR /&gt;My $0.02,&lt;BR /&gt;Jeff</description>
      <pubDate>Fri, 30 Apr 2004 17:02:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264728#M569293</guid>
      <dc:creator>Jeff Schussele</dc:creator>
      <dc:date>2004-04-30T17:02:14Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264729#M569294</link>
      <description>I guess I should clarify that a little more...&lt;BR /&gt;When disk I/O goes away a TOC is not the reaction - a hang will occur. IF a TOC finally happens - unlikely - then the failover will occur. &lt;BR /&gt;SO moral is only failures that cause TOCS or failures of monitored HW resources (hint LAN) will cause a failover.&lt;BR /&gt;NOW... there's nothing stopping you from setting up a monitor script that will watch the disk I/O &amp;amp; IF it sees a failure of *ONE* channel then cause a manual failover. But still - even a monitor script can't save you if BOTH channels fail because you'd still need to cause a TOC &amp;amp; you can't do that from the command line.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;Jeff</description>
      <pubDate>Fri, 30 Apr 2004 17:12:07 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264729#M569294</guid>
      <dc:creator>Jeff Schussele</dc:creator>
      <dc:date>2004-04-30T17:12:07Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264730#M569295</link>
      <description>We receive no error messages, as far as the O/S is concerned. We only receive the messages from Informix saying an Assert Failure I/O&lt;BR /&gt;and then I am told what chunk is going down&lt;BR /&gt;&lt;BR /&gt;So, we cannot get the O/S to go down different &lt;BR /&gt;paths is a controller go out, without spending &lt;BR /&gt;more money huh?&lt;BR /&gt;By the way we do not run Service Guard here.</description>
      <pubDate>Fri, 30 Apr 2004 17:25:24 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264730#M569295</guid>
      <dc:creator>ROSS HANSON</dc:creator>
      <dc:date>2004-04-30T17:25:24Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264731#M569296</link>
      <description>Ross,&lt;BR /&gt;&lt;BR /&gt;I am having a 11.0 server in the exact same configuration you have and it is running pretty smooth without any problems.It is one of our legacy applications which has not been migrated to 11i. &lt;BR /&gt;We ahve however had numerous problems with the VA and have stopped using it in favour of a XP1024. &lt;BR /&gt;Let me poke around my server and see if I can come up with something you can use.</description>
      <pubDate>Fri, 30 Apr 2004 17:30:19 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264731#M569296</guid>
      <dc:creator>hari jayaram_1</dc:creator>
      <dc:date>2004-04-30T17:30:19Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264732#M569297</link>
      <description>Sorry - when you said failover I thought you implied Service Guard.&lt;BR /&gt;You need to check the LVM setup of the VGs that contain that data:&lt;BR /&gt;vgdisplay -v /dev/vg_name&lt;BR /&gt;there *must* be an alternate link there for LVM to handle the link loss.&lt;BR /&gt;If not you need to vgextend that VG to use that extra link.&lt;BR /&gt;You should see something at the end of the output like:&lt;BR /&gt;/dev/dsk/c5t2d3&lt;BR /&gt;/dev/dsk/c9t2d3 Alternate Link&lt;BR /&gt;If you don't - you don't have that LVM protection.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;Jeff&lt;BR /&gt;</description>
      <pubDate>Fri, 30 Apr 2004 17:31:29 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264732#M569297</guid>
      <dc:creator>Jeff Schussele</dc:creator>
      <dc:date>2004-04-30T17:31:29Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264733#M569298</link>
      <description>Hi Ross,&lt;BR /&gt;&lt;BR /&gt;I saw this in one of the forum threads&lt;BR /&gt;&lt;A href="http://forums1.itrc.hp.com/service/forums/questionanswer.do?admit=716493758+1083364406421+28353475&amp;amp;threadId=215197" target="_blank"&gt;http://forums1.itrc.hp.com/service/forums/questionanswer.do?admit=716493758+1083364406421+28353475&amp;amp;threadId=215197&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;We have discovered the problem here.&lt;BR /&gt;&lt;BR /&gt;Linkdown-tmo was set to 60 and no_device_delay was set to 30.&lt;BR /&gt;&lt;BR /&gt;The combinations of these two delays caused the PVLinks to get in such a state that they never failed over.&lt;BR /&gt;&lt;BR /&gt;I waited for one hour until resetting the kernel parameters to defaults.&lt;BR /&gt;&lt;BR /&gt;All is now working properly.&lt;BR /&gt;&lt;BR /&gt;Thanks for everyone's input.</description>
      <pubDate>Fri, 30 Apr 2004 17:35:39 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264733#M569298</guid>
      <dc:creator>hari jayaram_1</dc:creator>
      <dc:date>2004-04-30T17:35:39Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264734#M569299</link>
      <description>When utilizing alternate paths the switch from primary to alternate is NOT instantaneous.  There can be a delay of several seconds while I/O times out on one path before it goes down the other path.  &lt;BR /&gt;&lt;BR /&gt;What you might want to look at is HP's Auto Path VA product.  I think that will give you more of the functionality that you want.&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://www.hp.com/products1/storage/products/disk_arrays/modular/autopath/index.html" target="_blank"&gt;http://www.hp.com/products1/storage/products/disk_arrays/modular/autopath/index.html&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Fri, 30 Apr 2004 17:36:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264734#M569299</guid>
      <dc:creator>Patrick Wallek</dc:creator>
      <dc:date>2004-04-30T17:36:01Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264735#M569300</link>
      <description>Yes Jeff,&lt;BR /&gt;We do have alternate links already. That is what brought up the whole thing with the fail over test.  As mentioned we have everything doubled, switches, disks, controllers ect...&lt;BR /&gt;and configured to supposible use alternate links to go to the other side if one should fail. But, our database has gone down twice now because when a san controller fails it does not go to the alternate link it just takes down the database. I thought with the redundency we have in place that it would "fail over"</description>
      <pubDate>Fri, 30 Apr 2004 17:41:54 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264735#M569300</guid>
      <dc:creator>ROSS HANSON</dc:creator>
      <dc:date>2004-04-30T17:41:54Z</dc:date>
    </item>
    <item>
      <title>Re: fail over test</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264736#M569301</link>
      <description>OK - then you need to look at the I/O timeout values for either the LV or PV and whether they are longer than what the SW will tolerate. &lt;BR /&gt;In LV &amp;amp; PV it's the -t value that sets it.&lt;BR /&gt;IF it's higher than the SW can tolerate then either the HW or SW value needs to change.&lt;BR /&gt;I believe about 90 seconds is the HW default &amp;amp; it may be the SW pukes &amp;amp; dies at that amount. BUT I would caution you to not decrease it unless you're sure loads will *never* cause that long of a delay - i.e. you may want to elongate the SW timeout before you shorten the HW timeout.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;Jeff</description>
      <pubDate>Fri, 30 Apr 2004 17:50:37 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/fail-over-test/m-p/3264736#M569301</guid>
      <dc:creator>Jeff Schussele</dc:creator>
      <dc:date>2004-04-30T17:50:37Z</dc:date>
    </item>
  </channel>
</rss>

