HPE GreenLake Administration
- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - HP-UX
- >
- Campus serviceguard config and lock disk question
Operating System - HP-UX
1833724
Members
2680
Online
110063
Solutions
Forums
Categories
Company
Local Language
back
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Forums
Discussions
Discussions
Discussions
Forums
Discussions
back
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Blogs
Information
Community
Resources
Community Language
Language
Forums
Blogs
Topic Options
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-15-2009 08:04 AM
05-15-2009 08:04 AM
Campus serviceguard config and lock disk question
Hi,
I have a 4 node SG cluster built as follows;
-2 remote sites. (site A and B)
-4 servers, 2 servers at each site. (1,2 at A and 3,4 at B)
-each server runs a package and fails over to a sister server at the alternate site.(1 fails over to 3 and 2 fails over to 4)
-we are using dual lock disk. each disk is in its own VG and is visible to all 4 nodes.
---A <--------> B
---1 <--------> 3
---2 <--------> 4
vglock1 vglock2
I have this nagging question about this dual lock disk setup. The managing SG guide (B3936-90065.pdf) at page 58 states:
"If one of the dual lock disks fails, ServiceGuard will detect this when it
carries out periodic checking, and it will write a message to the syslog
file. After the loss of one of the lock disks, the failure of a cluster node
could cause the cluster to go down."
Does this mean that if my B data center were to completely crash (instant loss of power, servers, telecom and disks go out). Would the nodes at data center A panic and then attempt to reform the cluster instead of staying online?
I have a 4 node SG cluster built as follows;
-2 remote sites. (site A and B)
-4 servers, 2 servers at each site. (1,2 at A and 3,4 at B)
-each server runs a package and fails over to a sister server at the alternate site.(1 fails over to 3 and 2 fails over to 4)
-we are using dual lock disk. each disk is in its own VG and is visible to all 4 nodes.
---A <--------> B
---1 <--------> 3
---2 <--------> 4
vglock1 vglock2
I have this nagging question about this dual lock disk setup. The managing SG guide (B3936-90065.pdf) at page 58 states:
"If one of the dual lock disks fails, ServiceGuard will detect this when it
carries out periodic checking, and it will write a message to the syslog
file. After the loss of one of the lock disks, the failure of a cluster node
could cause the cluster to go down."
Does this mean that if my B data center were to completely crash (instant loss of power, servers, telecom and disks go out). Would the nodes at data center A panic and then attempt to reform the cluster instead of staying online?
2 REPLIES 2
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
05-18-2009 12:44 AM
05-18-2009 12:44 AM
Re: Campus serviceguard config and lock disk question
A dual cluster lock is a compound lock which means under normal circumstances you have to get both locks. This is why in a normal cluster adding a second cluster lock does not give extra redundancy it makes the cluster less available since the failure of either lock disk would prevent the cluster getting a lock risking an entire cluster failure. i.e. doubling the risk of a cluster lock failure.
However, there is a difference between failing to get a lock due to detecting an error, and not being able to contact the cluster lock disk at all which is what happens when you have a site failure.
In the situation you describe, if you have a cluster lock fail at site B and then you lost the entire site B, the cluster nodes at site A would attempt to reform and after obtaining the cluster lock at site A would continue running. This is because the request to obtain cluster lock B would timeout rather than fail, and since the lock at A was obtained the cluster would form.
This contrasts to the situation where the cluster lock at site B fails, and then nodes 3 and 4 at site B fail at the same time but not the whole site. i.e. cluster lock B is still reachable but is failed and generates an I/O error rather than a timeout. In this situation nodes 1 and 2 at site A would also fail since they could reach the cluster lock at site B but could not obtain it even if they had obtained the lock at site A.
i.e. you are safe with your configuration and are using dual cluster locks correctly. Unlike many who think adding a second lock is the right thing to do to protect from cluster lock failures in a non campus cluster environment.
However, there is a difference between failing to get a lock due to detecting an error, and not being able to contact the cluster lock disk at all which is what happens when you have a site failure.
In the situation you describe, if you have a cluster lock fail at site B and then you lost the entire site B, the cluster nodes at site A would attempt to reform and after obtaining the cluster lock at site A would continue running. This is because the request to obtain cluster lock B would timeout rather than fail, and since the lock at A was obtained the cluster would form.
This contrasts to the situation where the cluster lock at site B fails, and then nodes 3 and 4 at site B fail at the same time but not the whole site. i.e. cluster lock B is still reachable but is failed and generates an I/O error rather than a timeout. In this situation nodes 1 and 2 at site A would also fail since they could reach the cluster lock at site B but could not obtain it even if they had obtained the lock at site A.
i.e. you are safe with your configuration and are using dual cluster locks correctly. Unlike many who think adding a second lock is the right thing to do to protect from cluster lock failures in a non campus cluster environment.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
04-17-2010 11:58 PM
04-17-2010 11:58 PM
Re: Campus serviceguard config and lock disk question
Thanks, closing thread
The opinions expressed above are the personal opinions of the authors, not of Hewlett Packard Enterprise. By using this site, you accept the Terms of Use and Rules of Participation.
Company
Events and news
Customer resources
© Copyright 2025 Hewlett Packard Enterprise Development LP