- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - OpenVMS
- >
- Re: %QMAN-W-LOWMEMORY
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Forums
Discussions
Discussions
Discussions
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:23 AM
05-03-2010 11:23 AM
VMS v7.2-2, AS800 512MB, uptime 328 01:25:12
"Suddenly" this morning, I'm receiving the above message regarding the queue manager. It is month-end, but I don't think it's any busier a month-end than others, job wise.
Thus far I have reduced the journal file size (DIAG 7) from 2M blocks to ~6K, added a pagefile and restarted some processes to make use of it.
Is there anything else I can do to help it right now? I'm waiting for a quiet time to restart the queue manager with a larger value for page file quota.
ES47's with gobs of memory are patiently waiting on the sidelines so no need to suggest I upgrade, we are progressing at glacial speeds in that direction.
Cheers,
Art
Solved! Go to Solution.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:44 AM
05-03-2010 11:44 AM
Re: %QMAN-W-LOWMEMORY
Clustered or non-clustered environment? Shared system disk or separate system disks?
I'm thinking along the lines of fixing the queue manager on each different node in turn.
However, I don't have access to verify everything at this instant.
- Bob Gezelter, http://www.rlgsc.com
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:46 AM
05-03-2010 11:46 AM
Re: %QMAN-W-LOWMEMORY
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:47 AM
05-03-2010 11:47 AM
Re: %QMAN-W-LOWMEMORY
I suspect you'll be restarting the queue manager.
Presuming it's not some other system-level leak.
And if I were suggesting an upgrade, memory would be a factor, but I'd also be looking at OpenVMS Alpha V8.3 or at OpenVMS I64 V8.3-1H1.
I'd recommend getting away from those long up-times, too; that's an inviting strategy on the surface, but one tends to leave your servers down-revision on patches. My preference is shorter up-times and preferably with some redundancies of servers, and with clustering or fail-over. Planning for failures and for upgrades, in other words, rather than trying to avoid that. (There's often no prize for long up-times, and there can be costs.)
Barring key applications that are locked into Alpha, I'd be moving to OpenVMS I64, too.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 01:19 PM
05-03-2010 01:19 PM
Re: %QMAN-W-LOWMEMORY
Our company will be changing hands "soon", we'll see what the new owners think of VMS.
Cheers,
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 02:03 PM
05-03-2010 02:03 PM
Re: %QMAN-W-LOWMEMORY
Have you checked for accumulating queue entries? Maybe a queue with /RETAIN?
I run a job daily to look for retained jobs, or jobs with procedures that don't match the most recent file.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 03:12 PM
05-03-2010 03:12 PM
Re: %QMAN-W-LOWMEMORY
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 07:23 PM
05-03-2010 07:23 PM
Re: %QMAN-W-LOWMEMORY
Looks like QMAN is not getting enough virtual memory to perform its work.
What is the JOB_LIMIT of the queue. JOB_LIMIT would indicate number of jobs
in the queue that execute in parallel.
Also, how many jobs in the queue do actually execute in parallel?
>> restart the queue manager with a larger value for page file quota.
Yes. This would give the queue manager some more virtual memory to work with.
Regards,
Murali
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 07:27 PM
05-03-2010 07:27 PM
Re: %QMAN-W-LOWMEMORY
Looks like fix is available for this problem.
From the kits VMS732_QMAN-V0100 and VMS82A_QMAN-V010
=============================================
Problem Description:
Systems which have a very large physical memory (more
then 10GB) and a large PAGEFILE.SYS sometimes produce the
following QMAN error messages.
%QMAN-W-LOWMEMORY, the queue manager process may require
more virtual memory than is currently available.
%QMAN-F-ALLOCMEM, error allocating virtual memory
-LIB-F-INSVIRMEM, insufficient virtual memory
%JBC-E-QMANDEL, unexpected queue manager process termination
=============================================
Regards,
Ketan
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 08:09 PM
05-03-2010 08:09 PM
Re: %QMAN-W-LOWMEMORY
Also,
>> VMS v7.2-2, AS800 512MB, uptime 328 01:25:12
Current version of VMS is V7.2-2.
You have to upgrage VMS to V73-2 or onwards in order to be able
to instal the QMAN patches which has fix related to "QMAN-W-LOWMEMORY"
problem.
Regards,
Murali
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 08:15 PM
05-03-2010 08:15 PM
Re: %QMAN-W-LOWMEMORY
Ok, The fix is not for V7.2-2. Did you check the queue manager's page file quota?
SDA> SET PROCESS/INDEX=
SDA> READ SYSDEF
SDA> FORMAT JIB
...
FFFFFFFF.81DC0C80 JIB$L_PGFLQUOTA 00000A00
FFFFFFFF.81DC0C84 JIB$L_PGFLCNT 00000000
...
Regards,
Ketan
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 08:26 PM
05-03-2010 08:26 PM
Re: %QMAN-W-LOWMEMORY
>> I'm waiting for a quiet time to restart the queue manager with a larger value for page file quota.
This can be the workaround for the problem.
Regards,
Ketan
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 10:27 PM
05-03-2010 10:27 PM
Re: %QMAN-W-LOWMEMORY
you might want to have a look at the problem analysis section of the QMAN-W-LOWMEMORY problem/solution in the patches referenced before:
5.2.5.3 Problem Analysis:
With a large PAGEFILE.SYS, the check for available memory in the Queue Manager may overflow a longword. The results of this overflow are the unnecessary LOWMEMORY warnings and the possible system crash.
As this patch does NOT seem to be available for V7.2-2, your only real 'workaround' seems to be a restart of the queue manager.
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:02 PM
05-03-2010 11:02 PM
Re: %QMAN-W-LOWMEMORY
I've seen this problem in 2004 and have the original IPMT text (from the problem escalation). It's some simple math problem (overflow) in the code.
Note that you can easily force a QUEUE_MANAGER process dump and have it restart automatically (queue commands will hang for as along as it takes to write the process dump, i.e. less than a minute):
$ MCR JBC$COMMAND
JBC$COMMAND> diag 4
%JBC-I-DIAGNOSTIC,
Log for playback = 0
Save old Journal files = 0
Log all requests = 0
Dump on error = 0
Checkpoint: State = 0, In-memory blocks = 100
PersAlpha CHAALP-E8.4 $
%%%%%%%%%%% OPCOM 4-MAY-2010 08:39:33.24 %%%%%%%%%%%
Message from user SYSTEM on CHAALP
%QMAN-F-DIAGNOSTIC, A request was made to dump the queue manager.
Note that increasing the amount of pagefile space is contraproductive in this case !
Look at PAGFILCNT of the QUEUE_MANAGER process with F$GETJPI("pid-of-queue_manager","PAGFILCNT"). If it's getting near 214 million, you may see that problem.
Volker.
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-03-2010 11:31 PM
05-03-2010 11:31 PM
Re: %QMAN-W-LOWMEMORY
>> It's some simple math problem (overflow) in the code.
Volker's right.
Systems with large physical memory and large pagefile.sys would
face this problem. The problem was with the check for available memory
in the queue manager which could overflow the longword boundary.
The "QMAN-W-LOWMEMORY" messages logged were as a result of this overflow.
>> Note that increasing the amount of pagefile space is contraproductive
>> in this case !
Yes. Even after increasing the pagefile space, you could see the same old
error messages again and may be even more number of times.
Intalling the QMAN patch would be the way to go forward. But for this you
have to upgrade the current version of OpenVMS on the system to V73-2 or
onwards.
Regards,
Murali
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 04:42 AM
05-04-2010 04:42 AM
Re: %QMAN-W-LOWMEMORY
in the queue that execute in parallel.
Also, how many jobs in the queue do actually execute in parallel?"
Well, as in most VMS systems/clusters, we have more than one queue! There are about 370 ... ~110 batch queues and the rest print queues. Impossible to say how many jobs execute in parallel at any given time.
"Ketan: Looks like fix is available for this problem."
Great, except these systems are v7.2-2 and aren't going to be upgraded.
"Murali: You have to upgrage VMS to V73-2 or onwards ..."
Not going to happen.
"Ketan: pagefile quota"
FFFFFFFF.812EF840 JIB$L_PGFLQUOTA 00009EB0
FFFFFFFF.812EF844 JIB$L_PGFLCNT 00000950
Which "matches" what I see at DCL:
$ write sys$output f$getjpi(20307833,"PGFLQUOTA") 649984 (%x9EB00)
$ write sys$output f$getjpi(20307833,"PAGFILCNT") 38144 (%x9500)
"Volker: With a large PAGEFILE.SYS ..."
The original single pagefile is 1,300,000 blocks and I added another 1,300,000 block one. That's not "large" is it?
"Volker: ...you can easily force a QUEUE_MANAGER process dump and have it restart automatically (queue commands will hang for as along as it takes to write the process dump."
What happens to currently executing / printing jobs? Evaporate, or also just hang until the mgr comes back? The points for that suggestion depend on the answer. ;-)
Cheers,
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:17 AM
05-04-2010 05:17 AM
Re: %QMAN-W-LOWMEMORY
Well the time was right, I did the DIAG 4. What followed was one of the longer 3 or 4 minutes of my life ... the queue manager restarted and went solid computable and the cluster "hung" ... not quite, as I was seeing OPCOM messages about users trying to login, timing out. But all of my commands entered stalled. It did finish up whatever it was doing and things are back to "normal".
One exception, I can't use the f$getjpi lexical to get the pagfilcnt and pgflquota:
$ pipe show sys | search sys$pipe queue
2030CD04 QUEUE_MANAGER HIB 9 1893 0 00:08:31.90 4434 3267
$ write sys$output f$getjpi(2030CD04,"PAGFILCNT")
%DCL-W-IVCHAR, invalid numeric value - check for invalid digits
\2030CD04\
$ write sys$output f$getjpi(2030CD04,"PGFLQUOTA")
%DCL-W-IVCHAR, invalid numeric value - check for invalid digits
\2030CD04\
I can't do this for any process. WTF?
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:22 AM
05-04-2010 05:22 AM
Solutionplease include the double-quotes around the process-id:
AXPVMS $ write sys$output f$getjpi(26600E13,"PAGFILCNT")
%DCL-W-IVCHAR, invalid numeric value - check for invalid digits
\26600E13\
AXPVMS $ write sys$output f$getjpi("26600E13","PAGFILCNT")
495136
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:25 AM
05-04-2010 05:25 AM
Re: %QMAN-W-LOWMEMORY
$ pipe show sys | search sys$pipe queue
2030CD04 QUEUE_MANAGER HIB 9 2805 0 00:08:32.74 4434 3267
$ write sys$output f$getjpi("2030CD04","PGFLQUOTA")
649984
$ write sys$output f$getjpi("2030CD04","PAGFILCNT")
592080
All's well again. ;-)
Cheers,
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:26 AM
05-04-2010 05:26 AM
Re: %QMAN-W-LOWMEMORY
With all due respect, you need to put the hexadecimal Process ID in quotes, to wit:
$ WRITE SYS$OUTPUT F4GETJPI("05CE","BIOCNT")
Otherwise, the DCL parsing does not identify the first parameter as a literal constant, it identifies it as the name of a DCL symbol (hint: a process ID of aced would otherwise be ambiguous).
- Bob Gezelter, http://www.rlgsc.com
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:39 AM
05-04-2010 05:39 AM
Re: %QMAN-W-LOWMEMORY
executing batch and print jobs continue unharmed, if you restart the QUEUE_MANAGER. On my standalone V8.3 test system, it took about 21 seconds to write the .DMP file - I at least verfified, that the command worked, before I posted it.
The time for writing the forced dump certainly depends on the virtual memory used by the QUEUE_MANAGER. Did you check the size and the creation/modification date of SYS$SYSTEM:QMAN$QUEUE_MANAGER.DMP ?
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 05:45 AM
05-04-2010 05:45 AM
Re: %QMAN-W-LOWMEMORY
>> Great, except these systems are v7.2-2 and aren't going to be upgraded.
Upgrading would have given you access to the patches which has the
fix for the problem.
As upgrading is not a option, other alternative is to use a workaround.
i.e. to increase the pagefile size. Again with this change also, the problem
might come back again. Its a tricky situation.
As of now, how frequently are you seeing the "QMAN-W-LOWMEMORY" messages?
Regards,
Murali
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 06:10 AM
05-04-2010 06:10 AM
Re: %QMAN-W-LOWMEMORY
Volker, the dump file only took 45 seconds to write, but for some reason the queue manager was working "very hard" when it restarted. Definately "paused" the cluster for several minutes while it COMputed something.
Thanks,
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 06:14 AM
05-04-2010 06:14 AM
Re: %QMAN-W-LOWMEMORY
$ write sys$output f$getjpi("2030CD04","PGFLQUOTA")
649984
$ write sys$output f$getjpi("2030CD04","PAGFILCNT")
592080
Cheers,
Art
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-04-2010 06:24 AM
05-04-2010 06:24 AM
Re: %QMAN-W-LOWMEMORY
>> since the queue manager reset 40 minutes ago, I have not seen the
>> message again. The queue manager's pagefile quota and count are
>> back to "normal"
Hmm. Looks like the workaround would be to restart the queue manager
at a silent time.
Increasing the pagefile quota might not be required as the problem may not be
related to that (i.e. problem is due to overflow of longword in the code)
Regards,
Murali