- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - HP-UX
- >
- ninode usage on 11.0
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-20-2000 01:46 PM
тАО11-20-2000 01:46 PM
ninode usage on 11.0
Could somebody confirm whether a high NINODE usage is indicative of problems? I have opinions from others here that high usage is linked to high CPU utilisation, and that NINODE should be increased. My understanding (and also that of some local HP consultants) is that NINODE is simply a cache that fills and frees as it goes. Are there any recommendations out there on how NINODE *should* be configured?
thanking you!
neil.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-20-2000 01:57 PM
тАО11-20-2000 01:57 PM
Re: ninode usage on 11.0
In and of itself, a high ninode table is not a problem. 'ninode' represents the number of slots in the inode table, and thus the maximum number of open inodes that can be in memory at any given time. The table is used as a cache memory with the most recent open inodes kept in memory. Since each unique open file has an open inode associated with it, the larger the number of unique open files, the larger ninode should be.
See this document for more related information:
http://docs.hp.com/hpux/onlinedocs/os/KCparams.OverviewAll.html
...JRF...
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-20-2000 02:02 PM
тАО11-20-2000 02:02 PM
Re: ninode usage on 11.0
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-20-2000 02:35 PM
тАО11-20-2000 02:35 PM
Re: ninode usage on 11.0
However, it is for HFS filesystems only and with 10.xx and now 11.0, VXFS is the filesystem of choice with only /stand left as (required) HFS filesystem. So ninode only needs to be large enough to accomodate all the unique files that might be opened at the same time in /stand plus NFS files and directories. It has nothing to do with CPU utilization.
Now if you have not changed SAM's formula for ninode, it is likely *way* too large (10,000 to 20,000). You can safely reduce this to a fixed value, perhaps 500 for a system with no NFS and all VXFS filesystems except /stand) to perhaps 4,000 for a busy NFS system. ninode is configured in the kernel paramters section.
Bill Hassell, sysadmin
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
тАО11-20-2000 11:44 PM
тАО11-20-2000 11:44 PM
Re: ninode usage on 11.0
if sar -v reports a full utilization of ninode table system performance is affected. This problem is clear when you run top or swapinfo and response time is too low, even minutes. In this situation cpu util in usr mode its anormaly high.
Configuring a high number of ninode solve this low performance.
DNLS ( drectory name lookup cache) reported by gpm seems to be in function of ninode.