<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Disk Queue Length in Operating System - OpenVMS</title>
    <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440649#M72339</link>
    <description>Wim,&lt;BR /&gt;&lt;BR /&gt;CLUSTER_CREDIT = 128 is the maximum value.&lt;BR /&gt;&lt;BR /&gt;Is anything unusual happening with the disk, pathes to the disk or HSG80 ? Mount-verification ? Path switches ?&lt;BR /&gt;&lt;BR /&gt;Check the FC counters with SDA&amp;gt; FC STDT&lt;BR /&gt;(QF seen or Seq TMO &amp;gt; 0 ?).&lt;BR /&gt;&lt;BR /&gt;Is there a specific job, which always starts at 06:00 ?&lt;BR /&gt;&lt;BR /&gt;Volker.</description>
    <pubDate>Fri, 10 Dec 2004 06:46:25 GMT</pubDate>
    <dc:creator>Volker Halle</dc:creator>
    <dc:date>2004-12-10T06:46:25Z</dc:date>
    <item>
      <title>Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440645#M72335</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;I found that 1 disk in my cluster had at one moment a queue length of 90. At 6:00 it was 30, at 6:02 it was 90 and at 6:04 it was 5.&lt;BR /&gt;&lt;BR /&gt;The event is repeated every day at the same time.&lt;BR /&gt;&lt;BR /&gt;The active image is dataserver (Sybase).&lt;BR /&gt;&lt;BR /&gt;Top hot files reports high Non Virtual QIO (about 30% of all IO, mainly writes). This indicates file system activity ?&lt;BR /&gt;&lt;BR /&gt;What can this be ?&lt;BR /&gt;(VMS 7.3, GS160, HSG80, MA8000, FDDI)&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Fri, 10 Dec 2004 03:45:10 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440645#M72335</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T03:45:10Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440646#M72336</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;why not try to collect some info about those IOs with my favourite tool: SDA.&lt;BR /&gt;&lt;BR /&gt;If this happens predictably at the same time every day, just submit a batch job to run at 06:00 and include the following commands:&lt;BR /&gt;&lt;BR /&gt;$ ANAL/SYS&lt;BR /&gt;SDA&amp;gt; SET OUT/NOINDEX file1.lis&lt;BR /&gt;SDA&amp;gt; SHOW DEV &lt;DISKNAME&gt;&lt;BR /&gt;SDA&amp;gt; EXIT&lt;BR /&gt;&lt;BR /&gt;This should give you the list of IOs in the queue for the device and you can find out, what they are alike, who issued them etc.&lt;BR /&gt;&lt;BR /&gt;Volker.&lt;/DISKNAME&gt;</description>
      <pubDate>Fri, 10 Dec 2004 04:57:04 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440646#M72336</guid>
      <dc:creator>Volker Halle</dc:creator>
      <dc:date>2004-12-10T04:57:04Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440647#M72337</link>
      <description>That's an option Volker. In TNG I found some extra info :&lt;BR /&gt;1) both cluster nodes have a peak at the same moment : 70 for 1 and 20 for the other.&lt;BR /&gt;2) thruput was about 4 MB/sec at the moment of the peak&lt;BR /&gt;3) there was also a peak of 35 in "credit waits"&lt;BR /&gt;&lt;BR /&gt;Wim&lt;BR /&gt;</description>
      <pubDate>Fri, 10 Dec 2004 06:26:11 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440647#M72337</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T06:26:11Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440648#M72338</link>
      <description>Correction : thruput was about 9 MB/sec on 1 node, about 3.5 on the other. Thus about at its maximum.&lt;BR /&gt;&lt;BR /&gt;cluster_credits is at 128. Too low ?&lt;BR /&gt;Since the disk is shadowed via FDDI : mscp_buffer is at 16384 and mscp_credits at 128.&lt;BR /&gt;&lt;BR /&gt;Wim</description>
      <pubDate>Fri, 10 Dec 2004 06:38:59 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440648#M72338</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T06:38:59Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440649#M72339</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;CLUSTER_CREDIT = 128 is the maximum value.&lt;BR /&gt;&lt;BR /&gt;Is anything unusual happening with the disk, pathes to the disk or HSG80 ? Mount-verification ? Path switches ?&lt;BR /&gt;&lt;BR /&gt;Check the FC counters with SDA&amp;gt; FC STDT&lt;BR /&gt;(QF seen or Seq TMO &amp;gt; 0 ?).&lt;BR /&gt;&lt;BR /&gt;Is there a specific job, which always starts at 06:00 ?&lt;BR /&gt;&lt;BR /&gt;Volker.</description>
      <pubDate>Fri, 10 Dec 2004 06:46:25 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440649#M72339</guid>
      <dc:creator>Volker Halle</dc:creator>
      <dc:date>2004-12-10T06:46:25Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440650#M72340</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;&lt;QUOTE&gt;&lt;BR /&gt;Since the disk is shadowed via FDDI : mscp_buffer is at 16384 and mscp_credits at 128.&lt;BR /&gt;&lt;/QUOTE&gt;&lt;BR /&gt;&lt;BR /&gt;Maybe a superfluous question, but since you use FDDI, did you set NISCS_MAX_PKTSZ to 4486?&lt;BR /&gt;&lt;BR /&gt;Proost.&lt;BR /&gt;&lt;BR /&gt;Have one on me.&lt;BR /&gt;&lt;BR /&gt;Jan</description>
      <pubDate>Fri, 10 Dec 2004 07:25:34 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440650#M72340</guid>
      <dc:creator>Jan van den Ende</dc:creator>
      <dc:date>2004-12-10T07:25:34Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440651#M72341</link>
      <description>Jan, yes it is at 4468 (Set by Johan Michiels, so, should be ok).&lt;BR /&gt;&lt;BR /&gt;But I did mon clu and added cr_waits.&lt;BR /&gt;There are a few thousand of them but all nodes I checked have them, even in other companies.&lt;BR /&gt;&lt;BR /&gt;Volker : nothing special active. Just an application peak.&lt;BR /&gt;&lt;BR /&gt;Other thing : the FDDI is shared with another cluster. This one had a thruput of about 2-4 MB at the moment of the problem.&lt;BR /&gt;&lt;BR /&gt;Wim</description>
      <pubDate>Fri, 10 Dec 2004 07:33:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440651#M72341</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T07:33:44Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440652#M72342</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;here is an article explaining non-virtual QIOs as reported by TNG/PSDC&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://h18000.www1.hp.com/support/asktima/operating_systems/CHAMP_SRC931006004627.html" target="_blank"&gt;http://h18000.www1.hp.com/support/asktima/operating_systems/CHAMP_SRC931006004627.html&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;In your case, this would point to 'database', probabyl doing Logical-IO to some of it's files. And if it is happening every day at 06:00, there must be some 'time-released' job in the application causing this.&lt;BR /&gt;&lt;BR /&gt;Did you check the FC counters ?&lt;BR /&gt;&lt;BR /&gt;Volker.</description>
      <pubDate>Fri, 10 Dec 2004 07:38:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440652#M72342</guid>
      <dc:creator>Volker Halle</dc:creator>
      <dc:date>2004-12-10T07:38:44Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440653#M72343</link>
      <description>Now I took a look at a wider interval.&lt;BR /&gt;&lt;BR /&gt;It seems that there are also peaks at other hours. And the peak at 6:00 has been lowered because the interval is 30 min instead of 2.&lt;BR /&gt;&lt;BR /&gt;Between 01:06 and 2:15 several disks have peak queues of 10-70. And at this moment backup is active reading about 15 MB/sec. The SCS traffic during this interval is almost 0. I also saw backup doing lots of IO but not doing any thruput (during 20 minutes) generating queue length of 70 continuously.&lt;BR /&gt;&lt;BR /&gt;Wim&lt;BR /&gt;&lt;BR /&gt;So I guess it is normal behaviour when reading or writing too fast.</description>
      <pubDate>Fri, 10 Dec 2004 07:54:05 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440653#M72343</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T07:54:05Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440654#M72344</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;are you running your BACKUP jobs from an account with a very high DIOLM ? You may be overloading your HSG80...&lt;BR /&gt;&lt;BR /&gt;The SDA&amp;gt; FC STDT/ALL counters would give an indication (QF seen).&lt;BR /&gt;&lt;BR /&gt;Volker.</description>
      <pubDate>Fri, 10 Dec 2004 08:03:04 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440654#M72344</guid>
      <dc:creator>Volker Halle</dc:creator>
      <dc:date>2004-12-10T08:03:04Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440655#M72345</link>
      <description>Volker,&lt;BR /&gt;&lt;BR /&gt;It is 7.3, so no /ALL.&lt;BR /&gt;&lt;BR /&gt;The DIOLM is indeed high (4096).&lt;BR /&gt;&lt;BR /&gt;But how do you tell that the HSG80 is overloaded ? If thruput stays high I don't have a problem with long queues. But I like to know why and who is doing it.&lt;BR /&gt;&lt;BR /&gt;Wim&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Fri, 10 Dec 2004 08:36:09 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440655#M72345</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-10T08:36:09Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440656#M72346</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;the only indication of an 'overloaded' HSG80 - that I know about - is seen in the FC counters: QF seen = 'Queue Full seen' or&lt;BR /&gt;Seq Tmo = sequential timeouts. You need to check the counters on all your FC pathes from the node to the HSG80:&lt;BR /&gt;&lt;BR /&gt;SDA&amp;gt; FC SHOW FGA0&lt;BR /&gt;SDA&amp;gt; FC STDT&lt;BR /&gt;SDA&amp;gt; FC SHOW FGB0&lt;BR /&gt;SDA&amp;gt; FC STDT&lt;BR /&gt;...&lt;BR /&gt;&lt;BR /&gt;If IOs to the disk/path/HSG80 are temporarily stalled, the queue length will stay high, but no IOs will be processed, they will all be pending.&lt;BR /&gt;&lt;BR /&gt;Volker.</description>
      <pubDate>Fri, 10 Dec 2004 08:54:05 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440656#M72346</guid>
      <dc:creator>Volker Halle</dc:creator>
      <dc:date>2004-12-10T08:54:05Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440657#M72347</link>
      <description>Have you run VTDPY on each controller in the pair to determine whether one or the other runs out of CPU or other resource. Maybe path switching to serve this unit from a controller which only has other units mostly idle at this time will improve recovery from this peak.</description>
      <pubDate>Fri, 10 Dec 2004 17:39:48 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440657#M72347</guid>
      <dc:creator>Tom O'Toole</dc:creator>
      <dc:date>2004-12-10T17:39:48Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440658#M72348</link>
      <description>Wim,&lt;BR /&gt;&lt;BR /&gt;A side note, not particularly on topic.&lt;BR /&gt;&lt;BR /&gt;Taken to its extreme, a short queue (queue length = 1) is not necessarily a good thing. For example, seek optimization algorithms cannot optimize without a queue to analyze. If the HSG has a problem with long queues, it is a BUG. The driver and the controller should jointly ensure that the queue length does not cause controller problems.&lt;BR /&gt;&lt;BR /&gt;Short of exhausting non-paged dynamic memory on the host OpenVMS system, and the performance problems caused by a dis-porportionate queue length on one particular device, there should be no problems with long queues.&lt;BR /&gt;&lt;BR /&gt;Of course, the individual performance of a particular process is a different question. Most analyses of overall system performance maximize the overall performance of the system.  &lt;BR /&gt;&lt;BR /&gt;Using DIOLM and other account quotas to manage the workload on the HSG is crude and imprecise at best, and self-defeating at worst. It is true that for a particular configuration, increasing quotas beyond a certain point yields dramatically decreasing benefits, but that is an entirely different issue.&lt;BR /&gt;&lt;BR /&gt;I hope that the above is helpful.&lt;BR /&gt;&lt;BR /&gt;- Bob Gezelter, &lt;A href="http://www.rlgsc.com" target="_blank"&gt;http://www.rlgsc.com&lt;/A&gt;</description>
      <pubDate>Sun, 12 Dec 2004 08:33:53 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440658#M72348</guid>
      <dc:creator>Robert Gezelter</dc:creator>
      <dc:date>2004-12-12T08:33:53Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440659#M72349</link>
      <description>Did some thinking at home. Maybe it is just that normally all write IO goes into the cache and is processed afterwards. But if the cache is full it may lead to queues. So, simply a longer interval of heavy load. I will check on Wednesday with the real data.&lt;BR /&gt;&lt;BR /&gt;Tom : yes running vtdpy could give some extra data but I don' have permission to come at 6:00.&lt;BR /&gt;&lt;BR /&gt;Wim@home</description>
      <pubDate>Sun, 12 Dec 2004 11:38:40 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440659#M72349</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2004-12-12T11:38:40Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440660#M72350</link>
      <description>It seems that I "fogot" this case.&lt;BR /&gt;&lt;BR /&gt;I now have a simular case. FAL was doing non-virtual IO with a thruput of 2 - 6 MByte per second. This during 1 hour, so several GB. But the application guys don't know why.&lt;BR /&gt;&lt;BR /&gt;Any idea's ?&lt;BR /&gt;&lt;BR /&gt;The cluster was rebooted and the problem is gone. But I would like to know what it was.&lt;BR /&gt;&lt;BR /&gt;Wim</description>
      <pubDate>Wed, 02 Nov 2005 06:01:56 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440660#M72350</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2005-11-02T06:01:56Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440661#M72351</link>
      <description>Wim, Is that really VMS 7.3 or VMS 7.3.x?&lt;BR /&gt;&lt;BR /&gt;What Volker is suspecting but maybe not articulating strongly enough is that there is a serious performance problem once you hit a QFULL condition (credits) on the fibre channel driver. One spike, and you'll slow down after that.&lt;BR /&gt;The recovery for that is way too gentle/slow.&lt;BR /&gt;This is addressed by VMS732_FIBRE_SCSI-V0700.&lt;BR /&gt;The fact that it gets better after reboot kind of confirms this.&lt;BR /&gt;That potential alternative to this is supposdly:&lt;BR /&gt;SDA&amp;gt; FC SET WTID/WWID=wwid-number/QFTIMED=1&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;About the 6 am/ hourly spikes possibly related to sybase... Coudl sybase be doing like what oracle calls a checkpoint? A great many IOs in one go to sync memory with disks?&lt;BR /&gt;Is there a knob in sybase to limit this? Like a 'max-write-io' perhaps?&lt;BR /&gt;&lt;BR /&gt;Now about Backup. The backup process settings  with DIOLM many thousands is 'old school'. It goes well beyond the point of diminishing returns. Please check the current process quota recommendation, or just set it back to 100 and try.&lt;BR /&gt;&lt;BR /&gt;met vriendelijke groetjes,&lt;BR /&gt;&lt;BR /&gt;Hein.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Wed, 02 Nov 2005 07:41:24 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440661#M72351</guid>
      <dc:creator>Hein van den Heuvel</dc:creator>
      <dc:date>2005-11-02T07:41:24Z</dc:date>
    </item>
    <item>
      <title>Re: Disk Queue Length</title>
      <link>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440662#M72352</link>
      <description>Hein,&lt;BR /&gt;&lt;BR /&gt;It is 7.3 patched until the patches of 12-mar-2003.&lt;BR /&gt;&lt;BR /&gt;Note that the reboot solved the problem that FAL does a lot of non-vir qio. But why ?&lt;BR /&gt;&lt;BR /&gt;The queue length problem is still present and imo caused by high activity (controller saturated). At that moment a backup is busy and also some big FTP. And some smaller Sybase dumps. And the controller is also used by another cluster that is doing backups combined with heavy DWH activity.&lt;BR /&gt;&lt;BR /&gt;I still have the peaks non-virtual qio but currently not when the queue length is high.&lt;BR /&gt;&lt;BR /&gt;And Sybase has no checkpoint activity.</description>
      <pubDate>Wed, 02 Nov 2005 08:12:33 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-openvms/disk-queue-length/m-p/3440662#M72352</guid>
      <dc:creator>Wim Van den Wyngaert</dc:creator>
      <dc:date>2005-11-02T08:12:33Z</dc:date>
    </item>
  </channel>
</rss>

