- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - HP-UX
- >
- Re: Understanding 11.23 memory metrics
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Forums
Discussions
Discussions
Discussions
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 10:34 AM
07-12-2006 10:34 AM
We have an itanium box running 11.23, 64GB of RAM. This box runs multiple heavily used PostgreSQL clusters. These DB clusters recently experienced multiple "out of memory" and "Deferred swap reservation failure" errors. We automatically sample glance's GBL_MEM_UTIL every 60 seconds, and the highest sampled value was 89.1% (i.e., ~7GB of RAM remaining). We do not think any process actually requested >7GB of RAM, though we could have overlooked something.
My question: Why did this occur so long before we got to 100% of RAM utilization according to GBL_MEM_UTIL?
Our dbc_max_pct is 10%, dbc_min_pct is 3%, and GBL_MEM_CACHE_UTIL is constantly at 10% (all the cache is being used).
We've off-loaded much of the memory demand present at the time of the errors. Here's the swapinfo -tam output from the moment:
$swapinfo -tam
Mb Mb Mb PCT START/ Mb
TYPE AVAIL USED FREE USED LIMIT RESERVE PRI NAME
dev 4096 0 4096 0% 0 - 1 /dev/vg00/lvol2
reserve - 4096 -4096
memory 65524 41731 23793 64%
total 69620 45827 23793 66% - 0 -
TIA.
Ed
Solved! Go to Solution.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 10:45 AM
07-12-2006 10:45 AM
SolutionAbout running out of memory. The first thing to verify is that the processes are allowed to grow as large as they want. maxdsiz and maxdsiz_64 control these fences. Start by increasing maxdsiz to 1700 megs and maxdsiz_64 to perhaps 4Gb or so. Then make sure there is no ulimit setting in a startup script that prevents a process from asking for the memory it needs.
Bill Hassell, sysadmin
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 10:55 AM
07-12-2006 10:55 AM
Re: Understanding 11.23 memory metrics
Bill Hassell, sysadmin
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 01:36 PM
07-12-2006 01:36 PM
Re: Understanding 11.23 memory metrics
Another way to look at this is that if you didn't have pseudoswap enabled ('swapmem_on' = 1), as evidenced by the "memory" line in the output of 'swapinfo', then you would need 64GB of device swap to utilize all of your physical memory. This is simply the rule of swap "reservation".
With pseudoswap enabled, 75% of your physical memory can be counted for the kernel's swap reservation requirement.
However, (.75 * 64) + 4 = 52 GB.
Thus, with only 4GB of device swap, you can only reserve ~ 52GB of process space, wasting about 12GB (64-52) of memory.
Regards!
...JRF...
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 01:57 PM
07-12-2006 01:57 PM
Re: Understanding 11.23 memory metrics
I'd really like to understand a little more about cache sizing strategy.
I recall that disk access is typically 1000 times more expensive than memory access. I hear and accept that a large OS cache causes the kernel to burn a lot of cycles managing the cache.
The relevant piece of hard data I seem to be missing is the performance curve of "disk access time" vs. "cache access time" (where cache access time includes all cache mgmnt time) as the size of the cache increases. If the cache management time for a 6GB cache is still faster than the disk access time, then really, really large OS caches would seem to make a lot of sense up to the point where CPU bottlenecks become an issue.
Are you saying the sweet spot in *that* curve is 1000M-1200MB? Can anyone point me to any published research for HPUX for that curve?
TIA.
Ed
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 02:10 PM
07-12-2006 02:10 PM
Re: Understanding 11.23 memory metrics
TIA.
Ed
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-12-2006 02:29 PM
07-12-2006 02:29 PM
Re: Understanding 11.23 memory metrics
a 100-300 deviation MiB from that value will not have huge impacts on performance. The benefits of larger UNIX buffer caches diminish because at much above this range the searches within the buffer cache begin to take more and more time. (One of the benefits of the 11.11 and up versions of HP-UX is that the buffer cache searches are hashed and are thus more efficient than their earlier counter parts. As a comparison the 10.20 sweet-spot was much nearer 400MiB -- although again, you had to measure. You should also note that most disk drives and certainly all most all disk arrays also have on-board caches which also buffer the i/o and this tends to diminish the role of very large buffer caches as well.
The only way to really determine the sweet spot for your platform, disks, and application is to measure but the values that have been suggested will be rather close to optimum.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-15-2006 09:03 AM
07-15-2006 09:03 AM
Re: Understanding 11.23 memory metrics
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-15-2006 09:21 AM
07-15-2006 09:21 AM
Re: Understanding 11.23 memory metrics
I think you will find that reducing dbc_max_pct to 3 or 4 % and allocating memory to the settings of the postgres database will be helpful.
I've found in general, if the database has memory settings like the Oracle SGA you will do better increasing the settings there and decreasing the OS buffer cache.
Database settings are a finer tool for the job, the OS being crude and general.
At this point, even though your Postgres database clusters are heavily used, your system does not seem to be under any stress.
Check vmstat and these tools.
http://www.hpux.ws/system.perf.sh
SEP
Owner of ISN Corporation
http://isnamerica.com
http://hpuxconsulting.com
Sponsor: http://hpux.ws
Twitter: http://twitter.com/hpuxlinux
Founder http://newdatacloud.com
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
07-15-2006 03:03 PM
07-15-2006 03:03 PM