<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: make_net_recovery dangling in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961623#M507844</link>
    <description>Glad You found it.&lt;BR /&gt;&lt;BR /&gt;btw, by 11.11 TCOE the scsictl command has an option to trigger the dreaded 'domain validation test', maybe think about running it twice daily, it should detect such errors. on the other hand, you could simply monitor the EMS event_log :)&lt;BR /&gt;&lt;BR /&gt;Have a nice evening.&lt;BR /&gt;</description>
    <pubDate>Wed, 14 Mar 2007 12:22:00 GMT</pubDate>
    <dc:creator>Florian Heigl (new acc)</dc:creator>
    <dc:date>2007-03-14T12:22:00Z</dc:date>
    <item>
      <title>make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961619#M507840</link>
      <description>&lt;!--!*#--&gt;Hi,&lt;BR /&gt;&lt;BR /&gt;I need to urgently patch an HP test box (given only limited time scope for installing GOLDQPK11i of Dec 2006 Support+).&lt;BR /&gt;Because I haven't applied this patch before (latest was June 2006) I wanted to make sure to have a recent Ignite image in advance.&lt;BR /&gt;I already meant to have done this by yesterday when I had a make_tape_recovery lingering for over 3 hours after that I deemed this attempt futile and killed the make_* procs smoothly, allowing them to remove their locks.&lt;BR /&gt;For this box I had to revert to make_tape_recovery because of incompatible releases of Ignite on this client (C.6.8.152) and the server (C.6.7.79) where in the recovery.log I was given the dissatisfying advice to either upgrade the server, or downgrade the client.&lt;BR /&gt;Upgrade of the server wasn't an option, as this one is part of three production servers igniting one another in a round-robin fashion.&lt;BR /&gt;This morning I had to dig through stock piles of HP Application CDs to at last arrive at one from March 2006 that luckilly bore the very release C.6.7.79 that I required to downgrade this client to align it with the server.&lt;BR /&gt;I must say, this is where Ignite really sucks.&lt;BR /&gt;On the one hand it requires you to have exactly same releases on clients and server, while on the other hand on the whole software.hp.com site I couldn't find a link that would lead me to a download for my, though obsolete, but required Ignite release.&lt;BR /&gt;After this pain was taken I started make_net_recovery as a batch job and am now experiencing the same idling.&lt;BR /&gt;&lt;BR /&gt;This is how the Ignite procs appear on the client in the process table now.&lt;BR /&gt;&lt;BR /&gt;# UNIX95= ps -x -o pid,ppid,stime,state,cpu,args -p 13896,13766&lt;BR /&gt;  PID  PPID    STIME S  C COMMAND&lt;BR /&gt;13896 13766 11:19:52 S  0 /usr/bin/sh /opt/ignite/data/scripts/make_sys_image -s local -L&lt;BR /&gt;13766 13765 11:19:49 S  0 /opt/ignite/bin/make_net_recovery -v -P s -s igux-server -x inc_entire=vg00 -x exclude=/tmp -x exclude=/var/adm/crash -x exclude=/var/tmp -x exclude=/var/spool/sw&lt;BR /&gt;&lt;BR /&gt;When I attach to both PIDs with tusc I can see them sleeping on a read().&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;# tusc -apf 13896&lt;BR /&gt;( Attached to process 13896 ("/usr/bin/sh /opt/ignite/data/scripts/make_sys_image -s local -L") [32-bit] )&lt;BR /&gt;[13896] read(3, 0x77ff46b8, 1024) .......................................... [sleeping]&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;In the latest/recovery.log on the server these are the last written lines.&lt;BR /&gt;&lt;BR /&gt;# tail /var/opt/ignite/clients/igux-client/recovery/latest/recovery.log &lt;BR /&gt;                                                /dev/vg01/lv_data02     /data02 0&lt;BR /&gt;&lt;BR /&gt;        ** 0 - The Volume Group or Filesystem is Not included in the&lt;BR /&gt;               System Recovery Archive&lt;BR /&gt;        ** 1 - The Volume Group or Filesystem is Partially included in the&lt;BR /&gt;               System Recovery Archive&lt;BR /&gt;        ** 2 - The Volume Group or Filesystem is Fully included in the&lt;BR /&gt;               System Recovery Archive&lt;BR /&gt;&lt;BR /&gt;       * Checking Versions of Recovery Tools&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;I can see that the NFS shares are mounted on the client.&lt;BR /&gt;&lt;BR /&gt;An lsof on the client's PIDs shows these open files.&lt;BR /&gt;&lt;BR /&gt;# lsof -nP -p 13896,13766      &lt;BR /&gt;COMMAND     PID USER   FD   TYPE     DEVICE SIZE/OFF  NODE NAME&lt;BR /&gt;make_net_ 13766 root  cwd    DIR     64,0x3     8192  5605 /root/scripts&lt;BR /&gt;make_net_ 13766 root  txt    REG     64,0x6   126976 70370 /opt (/dev/vg00/lvol6)&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7    36864 13819 /usr/lib/libnss_dns.1&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7    53248 14533 /usr/lib/libnss_files.1&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7    12794  6309 /usr/lib/tztab&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x6   544768 70336 /opt (/dev/vg00/lvol6)&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7   221184  5999 /usr/lib/libCsup_v2.2&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7   282624  5839 /usr/lib/libm.2&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7    12288  6028 /usr/lib/libisamstub.1&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7  1261568 13838 /usr/lib/libcl.2&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7  1822720  6084 /usr/lib/libc.2&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7    24576 30789 /usr/lib/libdld.2&lt;BR /&gt;make_net_ 13766 root  mem    REG     64,0x7   274432 30787 /usr/lib/dld.sl&lt;BR /&gt;make_net_ 13766 root    0u   REG     64,0x8     2411  4827 /var (/dev/vg00/lvol8)&lt;BR /&gt;make_net_ 13766 root    1w   REG     64,0x8      292 28518 /var (/dev/vg00/lvol8)&lt;BR /&gt;make_net_ 13766 root    2w   REG     64,0x8      292 28518 /var (/dev/vg00/lvol8)&lt;BR /&gt;make_net_ 13766 root    3wW  REG     64,0x8        0 28520 /var (/dev/vg00/lvol8)&lt;BR /&gt;make_net_ 13766 root    4u   REG     78,0x9     1609  7230 /var/opt/ignite/recovery/client_mnt/0x00306E08FFFF/recovery/2007-03-14,11:19/recovery.log&lt;BR /&gt;make_net_ 13766 root    5r  FIFO 0x491b1c48      0t0 88059 &lt;BR /&gt;make_sys_ 13896 root  cwd    DIR     64,0x3     8192  5605 /root/scripts&lt;BR /&gt;make_sys_ 13896 root  txt    REG     64,0x7   204800 30110 /usr/bin/rsh&lt;BR /&gt;make_sys_ 13896 root  mem    REG     64,0x7    24576 30789 /usr/lib/libdld.2&lt;BR /&gt;make_sys_ 13896 root  mem    REG     64,0x7  1822720  6084 /usr/lib/libc.2&lt;BR /&gt;make_sys_ 13896 root  mem    REG     64,0x7   274432 30787 /usr/lib/dld.sl&lt;BR /&gt;make_sys_ 13896 root    0u   REG     64,0x8     2411  4827 /var (/dev/vg00/lvol8)&lt;BR /&gt;make_sys_ 13896 root    1w  FIFO 0x491b1c48      0t0 88059 &lt;BR /&gt;make_sys_ 13896 root    2w  FIFO 0x491b1c48      0t0 88059 &lt;BR /&gt;make_sys_ 13896 root    3r  FIFO 0x48d237c8      0t0 88064 &lt;BR /&gt;make_sys_ 13896 root   28r   REG     64,0x6   103506 70301 /opt (/dev/vg00/lvol6)&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Does anyone see what is possibly causing the deadlock?&lt;BR /&gt;Was my batch job approach (which always has worked so far) wrong?&lt;BR /&gt;&lt;BR /&gt;Regards&lt;BR /&gt;Ralph</description>
      <pubDate>Wed, 14 Mar 2007 08:13:26 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961619#M507840</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-14T08:13:26Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961620#M507841</link>
      <description>Observed a hanging of an "lvlnboot -v vg00" that I executed in an tsm window.&lt;BR /&gt;Looks like I 've got a disk, controller or LVM problem on my Ignite client...&lt;BR /&gt;</description>
      <pubDate>Wed, 14 Mar 2007 09:25:17 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961620#M507841</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-14T09:25:17Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961621#M507842</link>
      <description>Ralph, &lt;BR /&gt;&lt;BR /&gt;I'm not sure this applies to your issue, but most of the time we track this down to stale (and hard) nfs mounts on the system.&lt;BR /&gt;&lt;BR /&gt;is there anything to be picked from dmesg?&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;florian</description>
      <pubDate>Wed, 14 Mar 2007 10:44:58 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961621#M507842</guid>
      <dc:creator>Florian Heigl (new acc)</dc:creator>
      <dc:date>2007-03-14T10:44:58Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961622#M507843</link>
      <description>&lt;!--!*#--&gt;Ouch, don't know why haven't focused earlier on vmunix messages to syslog.&lt;BR /&gt;Today a couple of hours ago there was a SCSI bus hang according to some entries in syslog.log on the Ignite client (see below).&lt;BR /&gt;(but after that nothing else from vmunix was logged)&lt;BR /&gt;&lt;BR /&gt;Though the lvlnboot returned after a while it looks like we've got a disk problem here.&lt;BR /&gt;&lt;BR /&gt;# lvlnboot -v vg00&lt;BR /&gt;lvlnboot: LIF information corrupt or not present on  "/dev/dsk/c2t2d0".&lt;BR /&gt;Use the "mkboot" command to initialize the LIF area.&lt;BR /&gt;Boot Definitions for Volume Group /dev/vg00:&lt;BR /&gt;Physical Volumes belonging in Root Volume Group:&lt;BR /&gt;        /dev/dsk/c2t2d0 (0/0/2/0.2.0)&lt;BR /&gt;        /dev/dsk/c1t2d0 (0/0/1/1.2.0) -- Boot Disk&lt;BR /&gt;Boot: lvol1     on:     /dev/dsk/c2t2d0&lt;BR /&gt;                        /dev/dsk/c1t2d0&lt;BR /&gt;Root: lvol3     on:     /dev/dsk/c2t2d0&lt;BR /&gt;                        /dev/dsk/c1t2d0&lt;BR /&gt;Swap: lvol2     on:     /dev/dsk/c2t2d0&lt;BR /&gt;                        /dev/dsk/c1t2d0&lt;BR /&gt;Dump: lvol2     on:     /dev/dsk/c1t2d0, 0&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;There are quite a few PEs stale by now.&lt;BR /&gt;&lt;BR /&gt;# pvdisplay -v $(vgdisplay -v vg00|awk '/PV Name/{print$NF}')|grep -c stale&lt;BR /&gt;1077&lt;BR /&gt;&lt;BR /&gt;Let's see if I can read from /dev/rdsk/c2t2d0&lt;BR /&gt;&lt;BR /&gt;# timex dd if=/dev/rdsk/c2t2d0 of=/dev/null bs=1024k count=100&lt;BR /&gt;&lt;BR /&gt;No it seems to hang.&lt;BR /&gt;This is pretty daft because only a few weeks ago one of the root disks of this box was replaced.&lt;BR /&gt;At least the 3 year warranty hasn't expired yet.&lt;BR /&gt;&lt;BR /&gt;I am sorry for having bothered you.&lt;BR /&gt;These hangings always must have a natural cause...&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix: SCSI: First party detected bus hang -- lbolt: 137026454, bus: 2&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   lbp-&amp;gt;state: 5060&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   lbp-&amp;gt;offset: f0&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   lbp-&amp;gt;uPhysScript: 81fba000&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  From most recent interrupt:&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   ISTAT: 01, SIST0: 00, SIST1: 00, DSTAT: 84, DSPS: 00000010&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  lsp: 0000000000000000&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  lbp-&amp;gt;owner: 00000000491a2e00&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   bp-&amp;gt;b_dev: bc022000&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   scb-&amp;gt;io_id: 21f5360&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   scb-&amp;gt;cdb: 28 00 00 00 00 01 00 00 10 00&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   lbolt_at_timeout: 137023254, lbolt_at_start: 137023254&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   lsp-&amp;gt;state: 10d&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  scratch_lsp: 00000000491a2e00&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  Pre-DSP script dump [ffffffff81fba030]:&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   78347500 0000000a 78350800 00000000&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   0e000004 81fba540 e0100004 81fba7d4&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:  Script dump [ffffffff81fba050]:&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   870b0000 81fba2d8 98080000 00000005&lt;BR /&gt;Mar 14 11:24:38 igux-client vmunix:   721a0000 00000000 98080000 00000001&lt;BR /&gt;Mar 14 11:24:39 igux-client vmunix: SCSI: Resetting SCSI -- lbolt: 137026554, bus: 2&lt;BR /&gt;Mar 14 11:24:39 igux-client vmunix: SCSI: Reset detected -- lbolt: 137026554, bus: 2&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix: SCSI: First party detected bus hang -- lbolt: 137030354, bus: 2&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   lbp-&amp;gt;state: 1060&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   lbp-&amp;gt;offset: f0&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   lbp-&amp;gt;uPhysScript: 81fba000&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  From most recent interrupt:&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   ISTAT: 01, SIST0: 00, SIST1: 00, DSTAT: 84, DSPS: 00000010&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  lsp: 0000000000000000&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  lbp-&amp;gt;owner: 000000004212ef00&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   bp-&amp;gt;b_dev: bc022000&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   scb-&amp;gt;io_id: 21f532b&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   scb-&amp;gt;cdb: 28 00 00 00 00 00 00 00 02 00&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   lbolt_at_timeout: 137027154, lbolt_at_start: 137027154&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   lsp-&amp;gt;state: 10d&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  scratch_lsp: 000000004212ef00&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  Pre-DSP script dump [ffffffff81fba030]:&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   78347200 0000000a 78350800 00000000&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   0e000004 81fba540 e0100004 81fba7c8&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:  Script dump [ffffffff81fba050]:&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   870b0000 81fba2d8 98080000 00000005&lt;BR /&gt;Mar 14 11:25:17 igux-client vmunix:   721a0000 00000000 98080000 00000001&lt;BR /&gt;Mar 14 11:25:18 igux-client vmunix: SCSI: Resetting SCSI -- lbolt: 137030454, bus: 2&lt;BR /&gt;Mar 14 11:25:18 igux-client vmunix: SCSI: Reset detected -- lbolt: 137030454, bus: 2&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix: SCSI: Request Timeout -- lbolt: 137036186, dev: bc022000&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   lbp-&amp;gt;state: 60&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   lbp-&amp;gt;offset: ffffffff&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   lbp-&amp;gt;uPhysScript: 81fba000&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  From most recent interrupt:&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   ISTAT: 22, SIST0: 00, SIST1: 04, DSTAT: 00, DSPS: 81fba540&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  lsp: 00000000491a2e00&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   bp-&amp;gt;b_dev: bc022000&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   scb-&amp;gt;io_id: 21f5360&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   scb-&amp;gt;cdb: 28 00 00 00 00 01 00 00 10 00&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   lbolt_at_timeout: 137033054, lbolt_at_start: 137033054&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   lsp-&amp;gt;state: 10d&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  lbp-&amp;gt;owner: 00000000491a2e00&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  scratch_lsp: 0000000000000000&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  Pre-DSP script dump [ffffffff81fba020]:&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   00000000 00000000 41020000 81fba290&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   78347e00 0000000a 78350800 00000000&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:  Script dump [ffffffff81fba040]:&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   0e000004 81fba540 e0100004 81fba7f8&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix:   870b0000 81fba2d8 0a000000 81fba548&lt;BR /&gt;Mar 14 11:26:15 igux-client vmunix: SCSI: Abort abandoned -- lbolt: 137036186, dev: bc022000, io_id: 21f5360, status: 200&lt;BR /&gt;</description>
      <pubDate>Wed, 14 Mar 2007 11:04:56 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961622#M507843</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-14T11:04:56Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961623#M507844</link>
      <description>Glad You found it.&lt;BR /&gt;&lt;BR /&gt;btw, by 11.11 TCOE the scsictl command has an option to trigger the dreaded 'domain validation test', maybe think about running it twice daily, it should detect such errors. on the other hand, you could simply monitor the EMS event_log :)&lt;BR /&gt;&lt;BR /&gt;Have a nice evening.&lt;BR /&gt;</description>
      <pubDate>Wed, 14 Mar 2007 12:22:00 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961623#M507844</guid>
      <dc:creator>Florian Heigl (new acc)</dc:creator>
      <dc:date>2007-03-14T12:22:00Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961624#M507845</link>
      <description>Hi Florian,&lt;BR /&gt;&lt;BR /&gt;thanks for reminding me of the scsictl option.&lt;BR /&gt;I will see if I can rig up a passive Nagios service check that could run twice or thrice a day.&lt;BR /&gt;Usually we have the EMS agent enabled on all our HP boxes.&lt;BR /&gt;But for some strange reason just on this box it has been disabled ;-)&lt;BR /&gt;I'm sure that otherwise I would have been noticed by email of the broken disk long before .&lt;BR /&gt;I have just filed an HW case via SCM, and asked HP for a replacement disk.&lt;BR /&gt;</description>
      <pubDate>Thu, 15 Mar 2007 03:39:31 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961624#M507845</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-15T03:39:31Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961625#M507846</link>
      <description>Sadly this disk type doesn't seem to support the domain value test.&lt;BR /&gt;I performed it here on the undamaged root disk which is of euqal brand and model.&lt;BR /&gt;&lt;BR /&gt;# scsictl -c domain_val /dev/rdsk/c1t2d0&lt;BR /&gt;domain_val: option is valid for only Ultra160 and later controllers.&lt;BR /&gt;&lt;BR /&gt;# diskinfo /dev/rdsk/c1t2d0&lt;BR /&gt;SCSI describe of /dev/rdsk/c1t2d0:&lt;BR /&gt;             vendor: HP 73.4G&lt;BR /&gt;         product id: ST373454LC      &lt;BR /&gt;               type: direct access&lt;BR /&gt;               size: 71687369 Kbytes&lt;BR /&gt;   bytes per sector: 512&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;# scsictl -c get_bus_parms -c get_target_parms /dev/rdsk/c1t2d0&lt;BR /&gt;&lt;BR /&gt;BUS LIMITS&lt;BR /&gt;----------&lt;BR /&gt;flags:          0x0&lt;BR /&gt;width:          16 bits (8 = Narrow; 16 = Wide)&lt;BR /&gt;req/ack offset: 31&lt;BR /&gt;xfer rate:      20000000&lt;BR /&gt;SPEED:          40 MB/s (Ultra Wide)&lt;BR /&gt;&lt;BR /&gt;BUS PARMS&lt;BR /&gt;---------&lt;BR /&gt;flags:          0x0&lt;BR /&gt;width:          16 bits (8 = Narrow; 16 = Wide)&lt;BR /&gt;req/ack offset: 31&lt;BR /&gt;xfer rate:      20000000&lt;BR /&gt;SPEED:          40 MB/s (Ultra Wide)&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;TARGET LIMITS&lt;BR /&gt;-------------&lt;BR /&gt;flags:          0x0&lt;BR /&gt;width:          16 bits (8 = Narrow; 16 = Wide)&lt;BR /&gt;req/ack offset: 31&lt;BR /&gt;xfer rate:      20000000&lt;BR /&gt;SPEED:          40 MB/s (Ultra Wide)&lt;BR /&gt;&lt;BR /&gt;NEGOTIATED TARGET VALUES&lt;BR /&gt;------------------------&lt;BR /&gt;flags:          0x0&lt;BR /&gt;width:          16 bits (8 = Narrow; 16 = Wide)&lt;BR /&gt;req/ack offset: 31&lt;BR /&gt;xfer rate:      20000000&lt;BR /&gt;SPEED:          40 MB/s (Ultra Wide)&lt;BR /&gt;</description>
      <pubDate>Thu, 15 Mar 2007 03:49:20 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961625#M507846</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-15T03:49:20Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961626#M507847</link>
      <description>&lt;!--!*#--&gt;Ugh, in preparation for the replacement disk I now have a dangling lvreduce process which also holds a lock sentinel in /etc/lvmconf/lvm_lock.&lt;BR /&gt;Since a pvdisplay on the defective disk still reported its status, though as unavailable,&lt;BR /&gt;I considered this disk according to the Cookbook as still being "attached",&lt;BR /&gt;and went on audaciously issueing&lt;BR /&gt;&lt;BR /&gt;# vgdisplay -v vg00|awk '/LV Name/{print$NF}'|xargs -n1 -i lvreduce -m 0 -A n {} /dev/dsk/c2t2d0&lt;BR /&gt;&lt;BR /&gt;However, I am glad to notice that this runs into a timeout after a couple of minutes.&lt;BR /&gt;Looks as if I would have to isssue the lvreduce commands separately line by line...&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Thu, 15 Mar 2007 04:39:38 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961626#M507847</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-15T04:39:38Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961627#M507848</link>
      <description>Hi, &lt;BR /&gt;&lt;BR /&gt;the timeout is mostly the PV timeout, defaults to 90 seconds, but it seems to grow with every lv you reduce.&lt;BR /&gt;&lt;BR /&gt;I forgot how to properly remove disks, I got lazy and sit through the timeouts nowadays.&lt;BR /&gt;&lt;BR /&gt;- Search for that 'when good disks go bad' pdf by hp, and/or look for instructions for using lvreduce with the 'pv key'.&lt;BR /&gt;That way the disk will be silently wiped from the lv's config.&lt;BR /&gt;&lt;BR /&gt;- And be sure to use lvreduce -A n to avoid the autobackup call, which will query all disks. Instead, after reducing all lv's and removing the pv from the vg, do vgcfgbackup /dev/vg00&lt;BR /&gt;</description>
      <pubDate>Thu, 15 Mar 2007 10:15:12 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961627#M507848</guid>
      <dc:creator>Florian Heigl (new acc)</dc:creator>
      <dc:date>2007-03-15T10:15:12Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961628#M507849</link>
      <description>About scsictl: &lt;BR /&gt;&lt;BR /&gt;it must be the controller, not the disk.&lt;BR /&gt;unfortunately &lt;BR /&gt;a) scsictl -c domain is not in HP-UX MC/EOE builds so we don't have it at work and I can't give you a list of controllers supporting it&lt;BR /&gt;b) not all (e.g. onboard) controllers are accompanied by EMS resource agents. I filed an enhancement request on this last year, but with no success.&lt;BR /&gt;&lt;BR /&gt;Maybe that's why you didn't get a notification.</description>
      <pubDate>Thu, 15 Mar 2007 21:08:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961628#M507849</guid>
      <dc:creator>Florian Heigl (new acc)</dc:creator>
      <dc:date>2007-03-15T21:08:01Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961629#M507850</link>
      <description>&amp;gt; About scsictl:&lt;BR /&gt;&amp;gt; it must be the controller, not the disk. &lt;BR /&gt;&lt;BR /&gt;Can't see how this scsictl command should applied to other than a disk device file.&lt;BR /&gt;&lt;BR /&gt;I can't figure a device file for the controller, and passing scsictl the HW path doesn't please it.&lt;BR /&gt;&lt;BR /&gt;e.g.&lt;BR /&gt;&lt;BR /&gt;# ioscan -knfCext_bus&lt;BR /&gt;Class     I  H/W Path  Driver S/W State   H/W Type     Description&lt;BR /&gt;=================================================================&lt;BR /&gt;ext_bus   0  0/0/1/0   c720 CLAIMED     INTERFACE    SCSI C896 Ultra Wide Single-Ended&lt;BR /&gt;ext_bus   1  0/0/1/1   c720 CLAIMED     INTERFACE    SCSI C896 Ultra Wide Single-Ended&lt;BR /&gt;ext_bus   2  0/0/2/0   c720 CLAIMED     INTERFACE    SCSI C87x Ultra Wide Single-Ended&lt;BR /&gt;ext_bus   3  0/0/2/1   c720 CLAIMED     INTERFACE    SCSI C87x Ultra Wide Single-Ended&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;# scsictl -c domain_val 0/0/2/0          &lt;BR /&gt;scsictl: Can't open device 0/0/2/0.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;# ioscan -knfH0/0/2/0&lt;BR /&gt;Class     I  H/W Path     Driver S/W State   H/W Type     Description&lt;BR /&gt;=====================================================================&lt;BR /&gt;ext_bus   2  0/0/2/0      c720  CLAIMED     INTERFACE    SCSI C87x Ultra Wide Single-Ended&lt;BR /&gt;target    5  0/0/2/0.0    tgt   CLAIMED     DEVICE       &lt;BR /&gt;disk      3  0/0/2/0.0.0  sdisk CLAIMED     DEVICE       SEAGATE ST318404LC&lt;BR /&gt;                         /dev/dsk/c2t0d0   /dev/rdsk/c2t0d0&lt;BR /&gt;target    6  0/0/2/0.2    tgt   CLAIMED     DEVICE       &lt;BR /&gt;disk      4  0/0/2/0.2.0  sdisk CLAIMED     DEVICE       HP 73.4GST373454LC&lt;BR /&gt;                         /dev/dsk/c2t2d0   /dev/rdsk/c2t2d0&lt;BR /&gt;target    7  0/0/2/0.7    tgt   CLAIMED     DEVICE       &lt;BR /&gt;ctl       2  0/0/2/0.7.0  sctl  CLAIMED     DEVICE       Initiator&lt;BR /&gt;                         /dev/rscsi/c2t7d0&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Ok, Initiator has a device file and sound plausible to me.&lt;BR /&gt;Unfortunately scsictl cannot communicate over it.&lt;BR /&gt;&lt;BR /&gt;# scsictl -c domain_val /dev/rscsi/c2t7d0&lt;BR /&gt;scsictl: Can't open device /dev/rscsi/c2t7d0.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Btw, the disk has been replaced by now.&lt;BR /&gt;And the originating make_*_recovery passed as usually without any errors.&lt;BR /&gt;</description>
      <pubDate>Fri, 16 Mar 2007 05:52:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961629#M507850</guid>
      <dc:creator>Ralph Grothe</dc:creator>
      <dc:date>2007-03-16T05:52:44Z</dc:date>
    </item>
    <item>
      <title>Re: make_net_recovery dangling</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961630#M507851</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;the domain validation test is controller initiated, but called by pointing at a disk :)&lt;BR /&gt;&lt;BR /&gt;The scsi uw controllers might actually be too old; I only know about the domain validation test from the errors we sometimes saw, and that was on C1010 controllers and beyond. (LSI-based dual channel u3w scsi)</description>
      <pubDate>Fri, 16 Mar 2007 07:52:02 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/make-net-recovery-dangling/m-p/3961630#M507851</guid>
      <dc:creator>Florian Heigl (new acc)</dc:creator>
      <dc:date>2007-03-16T07:52:02Z</dc:date>
    </item>
  </channel>
</rss>

