- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - HP-UX
- >
- LUN failover causing stale NFS mount points
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-28-2006 11:12 PM
тАО11-28-2006 11:12 PM
LUN failover causing stale NFS mount points
Im not sure exactly what category to place this under because it covers a host of issues.
We have quit a few installations that consist of an EVA5000 SAN hanging off two Itanium Unix Servers. The software we use works withing service guard and has soft NFS mounts to volume groups made from the disk groups presented by the SAN.
Unfortunately we occasionally experience power, switch or other failures that causes the LUN paths to failover. On the Unix boxes we use a combination of secure path and pvlinks, and when the LUN's failover the VG start using the alternate links automatically as they should. But this leaves stale NFS mounts and it causes many NFS write failures that end up using all of the CPU. We then have to move the package to the alternate node and restart the affected server.
Once the package has been moved, the mount points should have been removed from the second server but there still visible after entering the mount command.
Some snapshot messages from dmesg:
CPQswsp: Path c12t1d5 Failed (LUN 600508B40010344300007000014F0000 Controller P5849E1AAR5062 Array 50001FE15004A740 HBA fcd2)
CPQswsp: Path c12t1d6 Failed (LUN 600508B4001034430000700001550000 Controller P5849E1AAR5062 Array 50001FE15004A740 HBA fcd2)
CPQswsp: Availability for LUN 600508B4001034430000700001970000 changed to Reduced
CPQswsp: Availability for LUN 600508B40010344300007000019A0000 changed to Reduced
0/4/2/0: Device at device id 0xc0400 has disappeared from Name Server GPN_FT
(FCP type) response, or its 'Port World-Wide Name' has changed.
device id = loop id, for private loop devices
device id = nport ID, for fabric/public-loop devices
System won't be able to see LUNs behind this port.
CPQswsp: Availability for LUN 600508B4001034430000700001020000 changed to Alive
CPQswsp: Availability for LUN 600508B4001034430000700001090000 changed to Alive
NFS write failed for server may_ims4s: RPC: Timed out
NFS write failed for server may_ims4s: RPC: Timed out
As you can see, this is were it starts timing out.
So may questions are:
Is there a reason the NFS experiences such problems and is there a way to avoid it?
When it occurs, the only solution I have found online is to restart the server, does anyone know of a better solution?
Lastly, What are the most secure setting to use for NFS mounts in this situation?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-28-2006 11:47 PM
тАО11-28-2006 11:47 PM
Re: LUN failover causing stale NFS mount points
This is caused by your logical volume layout and general situation.
1) Your systems including disk should be able to endure short power flucations without any loss of service.
2) Take a look at lvdisplay -v output and lvm configuration to find the problem. If you have done something like strip a logical volume across LUNS, don't do that. Let the EVA handle disk coinfiguration, don't waste server cpu and i/o on that.
3) This problem should be avoidable by proper layout of lvm and configuration of nfs.
SEP
Owner of ISN Corporation
http://isnamerica.com
http://hpuxconsulting.com
Sponsor: http://hpux.ws
Twitter: http://twitter.com/hpuxlinux
Founder http://newdatacloud.com
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-29-2006 12:13 AM
тАО11-29-2006 12:13 AM
Re: LUN failover causing stale NFS mount points
If so, what version?
Also, what version of HP-UX and Serviceguard are you using?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-29-2006 12:22 AM
тАО11-29-2006 12:22 AM
Re: LUN failover causing stale NFS mount points
HPUXBaseOS B.11.23 HP-UX Base OS
T1905BA A.11.16.00 Serviceguard
As for the NFS, Im pretty sure that its NFS v3, the latest one supported by HPUX, but if somebody can let me know how to determine this preciously Ill post the results.
The LVM is correctly laid out on the system and I've check the lvdisplay command to confirm that nothing is stripped across the luns.
I presume that the system is correctly configured because this only occurs when theres a lun failover on the unix boxes.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-29-2006 01:20 AM
тАО11-29-2006 01:20 AM
Re: LUN failover causing stale NFS mount points
Serviceguard NFS Toolkit A.11.11.07 and A.11.23.06 Release Notes
http://docs.hp.com/en/ha.html#Highly%20Available%20NFS
While reading the A.11.23.06 release notes, symptoms that you are describing appear under the 'known problems and workaround' headings. For example:
Cause
HA/NFS uses relocatable IP addresses that migrate from a primary NFS server to an
adoptive node during a fail-over event. If you start AutoFS on an HA/NFS server after
Serviceguard is running, these relocatable IP addresses are included in AutoFS├в list of local IP
interfaces. AutoFS map entries referencing these relocatable addresses cause AutoFS to
create LOFS mounts to the exported local filesystems. These LOFS mounts can create
problems with Serviceguard, because HA/NFS does not unmount LOFS filesystems before a
package migration. AutoFS-managed LOFS mounts that are holding resources in the
exported filesystems, when Serviceguard tries to initiate a package failover, may render
Serviceguard unable to successfully unmount the local filesystems and migrate them to the
adoptive node.
Workaround
The workaround for this problem is to start automountd with the -L option. This can be
done by setting the AUTOMOUNTD_
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-29-2006 03:15 AM
тАО11-29-2006 03:15 AM