<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: bonding 10g rac failover in Operating System - Linux</title>
    <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417425#M86485</link>
    <description>hi rui amaral,&lt;BR /&gt;&lt;BR /&gt;it would be nice if you asign some points to the answers that helps you.&lt;BR /&gt;&lt;BR /&gt;thanks,&lt;BR /&gt;&lt;BR /&gt;johannes</description>
    <pubDate>Tue, 09 Nov 2004 15:52:43 GMT</pubDate>
    <dc:creator>Johannes Krackowizer_1</dc:creator>
    <dc:date>2004-11-09T15:52:43Z</dc:date>
    <item>
      <title>bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417418#M86478</link>
      <description>Hi all,&lt;BR /&gt;&lt;BR /&gt;Hope someone has a solution:&lt;BR /&gt;&lt;BR /&gt;I have the following:&lt;BR /&gt;&lt;BR /&gt;3 dl380 g3&lt;BR /&gt;6 nics per machine &lt;BR /&gt;RedHat AS 3 update 2&lt;BR /&gt;bcm5700-7.1.9e-1 &lt;BR /&gt;e1000-5.2.16b-1&lt;BR /&gt;bonding-1.0.4o-1&lt;BR /&gt;Oracle 10g RAC 10.1.0.2/10.1.0.3&lt;BR /&gt;NetApp FAS 270c&lt;BR /&gt;&lt;BR /&gt;I have all nics bonded successfully. And the system is stable. We are doing fault testing and are trying to mimic a complete loss to the bonds that manage the nfs storage or to the bond that manages the interconnect. We pull both cables. OS sees that the nics have faulted  but the bond still remains and the database still runs or hangs. When we put the cables back it then the database faults and the system reboots. &lt;BR /&gt;&lt;BR /&gt;The situation is we want the system to reboot when the all nics in a bond are gone. I know the document says that the bond will still be there when the nics fail but does someone have a way to make the bond fail in the scenario?&lt;BR /&gt;&lt;BR /&gt;Thanks.</description>
      <pubDate>Mon, 08 Nov 2004 11:22:56 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417418#M86478</guid>
      <dc:creator>Rui Amaral_1</dc:creator>
      <dc:date>2004-11-08T11:22:56Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417419#M86479</link>
      <description>This is not an Oracle RAC issue, its a bonding issue.&lt;BR /&gt;&lt;BR /&gt;Here is the doc I used to bond two Intel NIC cards. Note: If you are not using an Intel card or another card that supports bonding it won't work. You need to check to see if your NIC explicitly supports bonding.&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://www.redhat.com/docs/manuals/enterprise/RHEL-3-Manual/ref-guide/s1-modules-ethernet.html" target="_blank"&gt;http://www.redhat.com/docs/manuals/enterprise/RHEL-3-Manual/ref-guide/s1-modules-ethernet.html&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;It may say its bonded but still not work.&lt;BR /&gt;&lt;BR /&gt;SEP</description>
      <pubDate>Mon, 08 Nov 2004 15:10:10 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417419#M86479</guid>
      <dc:creator>Steven E. Protter</dc:creator>
      <dc:date>2004-11-08T15:10:10Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417420#M86480</link>
      <description>Sorry but yes I do know it's not a rac issue nor is it the nics (all nics support bonding - I checked months ago) as nic failover within the bond works fine - I have no problem with that at all.&lt;BR /&gt;&lt;BR /&gt;The scenario is this:&lt;BR /&gt;&lt;BR /&gt;Bond0 has 2 nics eth0 and eth1.&lt;BR /&gt;&lt;BR /&gt;Eth0 fails and bond now uses eth1.&lt;BR /&gt;Eth1 fails and there is no active nic in the bond. Bond0 does not fail at all which causes the database to hang.&lt;BR /&gt;&lt;BR /&gt;I have to manually bring down the bond which then causes the server to reboot (as expected within the rac environment).&lt;BR /&gt;&lt;BR /&gt;So instead of me manually bringing the bond down is there a parameter that I can pass to the bonding driver? or some sort of other parameter I can use to cause the bond to fail if all nics are dead?</description>
      <pubDate>Mon, 08 Nov 2004 15:28:34 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417420#M86480</guid>
      <dc:creator>Rui Amaral_1</dc:creator>
      <dc:date>2004-11-08T15:28:34Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417421#M86481</link>
      <description>hi Rui Amaral,&lt;BR /&gt;&lt;BR /&gt;i have no experince with nic bonding under linux but i know a workaround. try to run this script by cron every minute or whatever you want as interval.&lt;BR /&gt;&lt;BR /&gt;#!/bin/bash&lt;BR /&gt;&lt;BR /&gt;ifconfig eth0 2&amp;gt;/dev/null | grep UP &amp;gt;/dev/null 2&amp;gt;/dev/null&lt;BR /&gt;$returnvalue=`echo $?`&lt;BR /&gt;&lt;BR /&gt;if [ $returnvalue -ne 0 ]; then&lt;BR /&gt;&lt;BR /&gt; ifconfig eth1 2&amp;gt;/dev/null | grep UP &amp;gt;/dev/null 2&amp;gt;/dev/null&lt;BR /&gt; $returnvalue=`echo $?`&lt;BR /&gt;&lt;BR /&gt; if [ $returnvalue -ne 0 ]; then&lt;BR /&gt;  ifconfig bond0 down&lt;BR /&gt; fi&lt;BR /&gt;fi</description>
      <pubDate>Tue, 09 Nov 2004 07:10:27 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417421#M86481</guid>
      <dc:creator>Johannes Krackowizer_1</dc:creator>
      <dc:date>2004-11-09T07:10:27Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417422#M86482</link>
      <description>Thanks so much. That's what I was looking for!&lt;BR /&gt;&lt;BR /&gt;Here's what I got after reviewing your response:&lt;BR /&gt;&lt;BR /&gt;#!/bin/bash&lt;BR /&gt;&lt;BR /&gt;#this little script is to check for the bond and associated slaves being up.&lt;BR /&gt;&lt;BR /&gt;#bonding does not allow for the bond to fail even when both all slaces in the &lt;BR /&gt;#bond are down so this script will bring the bond down when all slaves are down.&lt;BR /&gt;&lt;BR /&gt;#adding the sleep command to start the script so that the service network restarrt&lt;BR /&gt;#does not cause the script to drop the bond&lt;BR /&gt;&lt;BR /&gt;sleep 60&lt;BR /&gt;echo "Slept for 1 minute"&lt;BR /&gt;&lt;BR /&gt;sleep 60&lt;BR /&gt;echo "Slept for 2 minutes"&lt;BR /&gt;&lt;BR /&gt;sleep 60&lt;BR /&gt;echo "Slept for 3 minutes"&lt;BR /&gt;&lt;BR /&gt;ifconfig bond2 2&amp;gt;/dev/null | grep UP  &amp;gt;/dev/null 2&amp;gt;/dev/null&lt;BR /&gt;valup=$?&lt;BR /&gt;echo $valup&lt;BR /&gt;&lt;BR /&gt;# a value of 0 means that the card is up and a value of 1 means that the card is down&lt;BR /&gt;&lt;BR /&gt;if [ $valup = 0 ]; then&lt;BR /&gt;echo 'Start test'&lt;BR /&gt;&lt;BR /&gt;until [ $valup != 0 ]&lt;BR /&gt;do&lt;BR /&gt;sleep 10&lt;BR /&gt;# this is to check for the first slave in the bond&lt;BR /&gt;        ifconfig eth4 2&amp;gt;/dev/null | grep UP  &amp;gt;/dev/null 2&amp;gt;/dev/null&lt;BR /&gt;        valeth4up=$?&lt;BR /&gt;        echo "A Value of 0 means the card is up"&lt;BR /&gt;        echo "State of eth4"&lt;BR /&gt;        echo $valeth4up&lt;BR /&gt;&lt;BR /&gt;#this is to check for the second slave in the bond&lt;BR /&gt;&lt;BR /&gt;        ifconfig eth5 2&amp;gt;/dev/null | grep UP  &amp;gt;/dev/null 2&amp;gt;/dev/null&lt;BR /&gt;        valeth5up=$?&lt;BR /&gt;         echo "State of eth5"&lt;BR /&gt;        echo $valeth5up&lt;BR /&gt;  &lt;BR /&gt;    if [ $valeth4up != 0 ]; then&lt;BR /&gt;       echo "First step :"&lt;BR /&gt;       echo $valeth4up &lt;BR /&gt; &lt;BR /&gt;       if [ $valeth5up != 0 ]; then&lt;BR /&gt;       echo "Second step might kill the bond"&lt;BR /&gt;       echo $valeth5up &lt;BR /&gt;       ifconfig bond2 down &lt;BR /&gt;       fi&lt;BR /&gt;    fi&lt;BR /&gt;done&lt;BR /&gt;fi&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Do you think this will do the trick? or am I missing something?</description>
      <pubDate>Tue, 09 Nov 2004 09:48:29 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417422#M86482</guid>
      <dc:creator>Rui Amaral_1</dc:creator>
      <dc:date>2004-11-09T09:48:29Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417423#M86483</link>
      <description>hi rui amaral,&lt;BR /&gt;&lt;BR /&gt;i think the script should work but i have some little things to say about the script:&lt;BR /&gt;&lt;BR /&gt;1) let the script run by cron so you don't need to sleep for 180 seconds. add the following line to /etc/crontab:&lt;BR /&gt;*/3 * * * * root /path/scriptname&lt;BR /&gt;&lt;BR /&gt;2) don't echo anything in the script because if the script echos anything an e-mail will be generated every 3 minutes. echo only if both nic's fail so you get an e-mail if the server restarts.&lt;BR /&gt;&lt;BR /&gt;3) why don't you do a 'shutdown -r now' when both nic's fail instead of 'ifconfig bond2 down'?&lt;BR /&gt;&lt;BR /&gt;best regards,&lt;BR /&gt;&lt;BR /&gt;johannes</description>
      <pubDate>Tue, 09 Nov 2004 15:04:26 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417423#M86483</guid>
      <dc:creator>Johannes Krackowizer_1</dc:creator>
      <dc:date>2004-11-09T15:04:26Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417424#M86484</link>
      <description>Ah great. Thanks. All good points that I will implement except for point 3.&lt;BR /&gt;&lt;BR /&gt;I just do the ifconfig down because I let the oracle cluster software determine when the node reboots so that the nodes can do a quick reconfig on their own before the problem one comes down.</description>
      <pubDate>Tue, 09 Nov 2004 15:09:47 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417424#M86484</guid>
      <dc:creator>Rui Amaral_1</dc:creator>
      <dc:date>2004-11-09T15:09:47Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417425#M86485</link>
      <description>hi rui amaral,&lt;BR /&gt;&lt;BR /&gt;it would be nice if you asign some points to the answers that helps you.&lt;BR /&gt;&lt;BR /&gt;thanks,&lt;BR /&gt;&lt;BR /&gt;johannes</description>
      <pubDate>Tue, 09 Nov 2004 15:52:43 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417425#M86485</guid>
      <dc:creator>Johannes Krackowizer_1</dc:creator>
      <dc:date>2004-11-09T15:52:43Z</dc:date>
    </item>
    <item>
      <title>Re: bonding 10g rac failover</title>
      <link>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417426#M86486</link>
      <description>Oh Yes. Sorry. I haven't gotten then hang of this set up yet.</description>
      <pubDate>Tue, 09 Nov 2004 16:01:41 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/bonding-10g-rac-failover/m-p/3417426#M86486</guid>
      <dc:creator>Rui Amaral_1</dc:creator>
      <dc:date>2004-11-09T16:01:41Z</dc:date>
    </item>
  </channel>
</rss>

