- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - OpenVMS
- >
- Re: Manual Que Manager & Print Que Cluster Failove...
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО05-30-2008 01:17 PM
тАО05-30-2008 01:17 PM
As part of a scheduled cluster node shutdown, we'd like to force the Que Manager and the telnetsymbiont processes to migrate to the remaining node (before the actual shutdown, where auto-failover would happen). (We presently have the Que Manger configured to run on any node, & the print queues are all defined with /AUTOSTART=(an explicit node list).)
How can I congifure for this setup?
What command(s) would I use to initiate the failover?
TIA
Solved! Go to Solution.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО05-30-2008 01:30 PM
тАО05-30-2008 01:30 PM
SolutionYou could and probably do use node-specific queues, stop the queue(s) of interest, and then merge the contents (eg: ASSIGN /MERGE ) to a queue on another host.
There are a gazillion tools around which skim the queues, and rebuild or reassign or otherwise manage the queues. (The FIXQUE.COM tool on the Freeware is a gonzo DCL tool that extracts all sorts of goodies out of the queue database, useful for rebuilding the queue database or as a launching board for other magicks.)
The direct command is STOP /QUEUE /ON_NODE ..., if you want to go that route. Read the DCL command reference documentation and the aforementioned System Manager's manual for details, and do expect active jobs will end up getting restarted.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО05-30-2008 04:33 PM
тАО05-30-2008 04:33 PM
Re: Manual Que Manager & Print Que Cluster Failover
Assuming /autostart=(tom::,jerry::) and you want to force all autostart queues to jerry.
JERRY$ enable autostart
TOM$ disable autostart
Or from any node:
$ enable autostart/queues/on_node=jerry::
$ disable autostart/queues/on_node=tom::
Examine sys$system:shutdown.com, and search for "disable". You will see that you can control when autostart queues are disabled.
That will take care of the queues with /autostart. For generic batch queues with execution queues on specific nodes, it is best (my opinion) to use stop/queue/next to allow currently executing jobs to continue, while preventing new jobs from starting on the specific execution queue.
Example:
$ init queue/batch batch$tom /on=tom:: ...
$ init queue/batch batch$jerry /on=jerry:: ...
$ init queue/batch/generic=(batch$tom,batch$jerry)
When ready to shutdown tom:
$ stop/queue/next batch$tom
I am not sure exactly what stop/queues/on_node does, but is it is more like stop/queue/reset, i.e. currently executing jobs are aborted. If your goal is to allow as many batch jobs to complete without them being aborted, then using stop/queues/on_node isn't the appropriate tool.
I don't know of any single command to do a stop/queue/next for all execution queues on a specific node. A command procedure using f$getqui could scan through all the non-autostart execution queues, find the execution node, and issue stop/queue/next commands.
Jon
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО06-02-2008 05:37 AM
тАО06-02-2008 05:37 AM
Re: Manual Que Manager & Print Que Cluster Failover
When we shut a node, we first do disable autostart/queue to get the failover. We don't wait for completion of the running jobs (we check it manually before the shutdown). All startup/shutdown is done in non-autostart queues.
When a node crashes, failover is automatic. But we signal all jobs in error (were busy at the moment of the crash). Of course we have /retain=error on all queues.
Wim
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО06-02-2008 08:54 AM
тАО06-02-2008 08:54 AM
Re: Manual Que Manager & Print Que Cluster Failover
$ start/queue/manager/on=(STAY_UP_NODE,*)
To move the telnet symbiont processes you must move the queues. Since they are all /AUTOSTART then use:
$ disable /autostart/on_node=GOING_DOWN_NODE
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО06-02-2008 11:28 PM
тАО06-02-2008 11:28 PM
Re: Manual Que Manager & Print Que Cluster Failover
Something often forgotten in interbuilding clusters is that spoolfiles must be on common disk (set dev/spool=common_dev, lpd printcap file).
Fwiw
Wim