- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - OpenVMS
- >
- Re: Continuation of earlier QMAN thread - question...
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 08:22 AM
тАО03-01-2009 08:22 AM
However...
1. We have not placed any restrictions on which nodes can host the queue manager, i.e. did not specify any "/ON=
Saturday night I shutdown the Itanium cluster members but not the Alpha. When the blades came back up, all of the batch/print queue entries were gone on the Itaniums. In addition, on the Alpha, teh couple of batch queues which run on that system were there, however all of the jobs were in a starting state.
My initial thought was that somehow the queue manager had failed over to the Alpha when the Itanium systems shutdown, and had not failed back.
Since the Alpha needed booting anyway, I shut it down, and rebooted. There were many messages indicating that the jobmanager had failed over to ECOM (one of the blades), however the jobs which were in a starting state were still in a starting state, and even worse. All of the queues hosted on the itanium blades were gone.
To cleanup, I issued a "stop /queue/manager/cluster", followed by a "start /queue /manager/on=
Question 1:
If I issue a "start /queue /manager /On=(Node1::,Node2::)", will this implicitly exclude Node3:: and Node4:: from taking part in any failover (this is the behaviour I want)
Question 2:
Should this command be executed every time the systems boot? Should it be executed on all nodes?
Dave.
Solved! Go to Solution.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 08:36 AM
тАО03-01-2009 08:36 AM
Re: Continuation of earlier QMAN thread - questions
does SHOW QUE/MANA/FULL show the same Database Location on all nodes in the cluster ? And is this location specified as a UNIQUE device and directory ? Using SYS$COMMON or any reference indirectly pointing to SYS$SYSDEVICE is incorrect !
$ START/QUE/MANA/ON=(Node1::,Node2::) excludes any nodes other than Node1 and Node2 from ever running the QUEUE_MANAGER process.
This command only needs to be executed ONCE. The information will be stored in QMAN$MASTER.DAT - this file also must be UNIQUE within the cluster, if you have a shared QMAN database.
The QUEUE_MANAGER does not automatically fail back to Node1, if Node1 gets rebooted. If it's running on Node2 and Node1 is up, you can use the command START/QUE/MANAGER to force failover to the Node1.
See the help text for $ HELP START/QUE/MANA/ON
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 08:47 AM
тАО03-01-2009 08:47 AM
Re: Continuation of earlier QMAN thread - questions
Here is the output of the "show/queue/manager/full", executed within SYSMAN.
BUD$SYSTEM>> mc sysman set env /clust
%SYSMAN-I-ENV, current command environment:
Clusterwide on local cluster
Username SYSTEM will be used on nonlocal nodes
SYSMAN> do show queu/mana/full
%SYSMAN-I-OUTPUT, command execution on node BUD
Master file: DSA101:[VMS$COMMON.SYSEXE]QMAN$MASTER.DAT;
Queue manager SYS$QUEUE_MANAGER, running, on BUD::
/ON=(BUD)
Database location: SYS$COMMON:[SYSEXE]
%SYSMAN-I-OUTPUT, command execution on node CITIUS
Master file: DSA101:[VMS$COMMON.SYSEXE]QMAN$MASTER.DAT;
Queue manager SYS$QUEUE_MANAGER, running, on BUD::
/ON=(BUD)
Database location: SYS$COMMON:[SYSEXE]
%SYSMAN-I-OUTPUT, command execution on node ECOM
Master file: DSA101:[VMS$COMMON.SYSEXE]QMAN$MASTER.DAT;
Queue manager SYS$QUEUE_MANAGER, running, on BUD::
/ON=(BUD)
Database location: SYS$COMMON:[SYSEXE]
%SYSMAN-I-OUTPUT, command execution on node SPEEDY
Master file: DSA101:[VMS$COMMON.SYSEXE]QMAN$MASTER.DAT;
Queue manager SYS$QUEUE_MANAGER, running, on BUD::
/ON=(BUD)
Database location: SYS$COMMON:[SYSEXE]
SYSMAN>
the logical QMAN$MASTER is defined as DSA101:[VMS$COMMON.SYSEXE]. DSA101 is a common cluster disk (and also the Itanium System Disk). It is mounted on the Alpha, in Sylogicals.com.
I am a little concerned that all system see the DB location as "SYS$COMMON:[SYSEXE]"
Dave
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 08:51 AM
тАО03-01-2009 08:51 AM
Re: Continuation of earlier QMAN thread - questions
And yes, use /ON.
For now, look to the restart to resolve this. And once you're on the SAN or otherwise with common disks (system disk or otherwise) for the cluster core files, to MSCPMOUNT or such, to keep the disks online on all nodes.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 08:54 AM
тАО03-01-2009 08:54 AM
Solution
I am a little concerned that all system see the DB location as "SYS$COMMON:[SYSEXE]"
That EXACTLY is your problem ! SYS$COMMON on Alpha cannot be the same as SYS$COMMON on Itanium !
You have to once stop the queue-manager cluster-wide and restart it with specifying a unique QMAN db location:
$ STOP/QUE/MANA/CLUSTER
$ START/QUE/MANA DISK$itanium:[VMS$COMMON.SYSEXE]
(assuming your *.QMAN* files are to be kept on the Itanium system disk.
Note that ALL queues will be stopped, if you issue the first command !
Volker.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 12:55 PM
тАО03-01-2009 12:55 PM
Re: Continuation of earlier QMAN thread - questions
This is exactly the problem I tried to warn you about in your last thread. Please reread my responses.
Move the queue manager files off all system disks, and make sure the logical names are IDENTICAL on all nodes. If you can see any differences in SHOW LOGICAL QMAN$MASTER executed on any node in the cluster, it's wrong.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО03-01-2009 05:07 PM
тАО03-01-2009 05:07 PM
Re: Continuation of earlier QMAN thread - questions
John, your responses were gratefully received and fully read. Unfortunately I was not in a position to do the complete move to a Common, NON-SYSTEM disk at this time so we had to make the more important system disk "authoratative", at least with respect to the Queue files.
The Logicals were all correctly defined in SYLOGICALS, however what I missed was to explicitly specify the location of the Queue Database when I started up the queue manager.
This was what Volker pointed out. Thanks.
As always, every day is a learning experience on this forum.
Dave.