- Community Home
- >
- Servers and Operating Systems
- >
- Operating System - OpenVMS
- >
- Re: Open VMS 7.3-2, a problem in the quorum disk
-
- Forums
-
- Advancing Life & Work
- Advantage EX
- Alliances
- Around the Storage Block
- HPE Ezmeral: Uncut
- OEM Solutions
- Servers & Systems: The Right Compute
- Tech Insights
- The Cloud Experience Everywhere
- HPE Blog, Austria, Germany & Switzerland
- Blog HPE, France
- HPE Blog, Italy
- HPE Blog, Japan
- HPE Blog, Middle East
- HPE Blog, Latin America
- HPE Blog, Russia
- HPE Blog, Saudi Arabia
- HPE Blog, South Africa
- HPE Blog, UK & Ireland
-
Blogs
- Advancing Life & Work
- Advantage EX
- Alliances
- Around the Storage Block
- HPE Blog, Latin America
- HPE Blog, Middle East
- HPE Blog, Saudi Arabia
- HPE Blog, South Africa
- HPE Blog, UK & Ireland
- HPE Ezmeral: Uncut
- OEM Solutions
- Servers & Systems: The Right Compute
- Tech Insights
- The Cloud Experience Everywhere
-
Information
- Community
- Welcome
- Getting Started
- FAQ
- Ranking Overview
- Rules of Participation
- Tips and Tricks
- Resources
- Announcements
- Email us
- Feedback
- Information Libraries
- Integrated Systems
- Networking
- Servers
- Storage
- Other HPE Sites
- Support Center
- Aruba Airheads Community
- Enterprise.nxt
- HPE Dev Community
- Cloud28+ Community
- Marketplace
-
Forums
-
Blogs
-
Information
-
English
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Email to a Friend
- Report Inappropriate Content
09-06-2012 01:03 AM - edited 09-06-2012 02:52 AM
09-06-2012 01:03 AM - edited 09-06-2012 02:52 AM
Open VMS 7.3-2, a problem in the quorum disk
2 nodes cluster, 2XAlpha DS25, MSA1000 by 2XHBA/server. the log:
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:39.68 %%%%%%%%%%% (from node SMIQ12 at 5-SEP-2012 08:38:13.09)
08:38:13.09 Node SMIQ12 (csid 00010002) timed-out operation to quorum disk
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:39.68 %%%%%%%%%%% (from node SMIQ12 at 5-SEP-2012 08:38:13.09)
08:38:13.09 Node SMIQ12 (csid 00010002) lost "connection" to quorum disk
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:40.18 %%%%%%%%%%%
08:38:40.18 Node SMIQ11 (csid 00010001) timed-out operation to quorum disk
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:40.18 %%%%%%%%%%%
08:38:40.18 Node SMIQ11 (csid 00010001) lost "connection" to quorum disk
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:40.18 %%%%%%%%%%%
08:38:40.18 Node SMIQ11 (csid 00010001) proposed modification of quorum or quorum disk membership
%%%%%%%%%%% OPCOM 5-SEP-2012 08:38:40.18 %%%%%%%%%%%
08:38:40.18 Node SMIQ11 (csid 00010001) completed VMScluster state transition
please, what's behind that?
thanks in advance.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Email to a Friend
- Report Inappropriate Content
09-06-2012 01:45 PM
09-06-2012 01:45 PM
Re: Open VMS 7.3-2, a problem in the quorum disk
Looks like you lost the quorum disk and the cluster did a proper state transistion. Is the actual disk drive still visible/available to the system?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Email to a Friend
- Report Inappropriate Content
09-06-2012 03:13 PM
09-06-2012 03:13 PM
Re: Open VMS 7.3-2, a problem in the quorum disk
Both your nodes are "Quorum disk watchers". That means they poll the quorum disk every QDSKINTERVAL seconds by sending a WRITE I/O. If enough I/Os time out, the quorum disk connection is declared "lost".
There are many possible reasons for this, the most common is a BACKUP involving the quorum disk or something common to it (adapter, bus, controller, etc...), saturating the resource and blocking the quorum disk polls. In that case the connection is typically reestablished. In your case the connection appears NOT to have been reestablished. The cluster has completed a state transition, kicking out the quorum disk. That suggest the disk is no longer visible to the nodes. Check the I/O paths to the disk and any physical hardware involved.
The cluster will continue to run, but if one of the nodes is lost, the remaining node will hang, waiting for more votes (which is presumably undesirable, since you have a quorum disk).
Recovery will depend on the results of the investigation as to what's happend to the disk.
Hewlett Packard Enterprise International
- Communities
- HPE Blogs and Forum
© Copyright 2021 Hewlett Packard Enterprise Development LP