- Community Home
- >
- Servers and Operating Systems
- >
- Operating Systems
- >
- Operating System - HP-UX
- >
- Re: IO wait
Categories
Company
Local Language
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Forums
Discussions
Discussions
Forums
Discussions
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Community
Resources
Forums
Blogs
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-12-2002 12:47 PM
10-12-2002 12:47 PM
IO wait
On my database server we see 40 - 50 %io wait reported by sar. Using glanceplus we verified that everything is normal. What is causing the IO wait? Could it be database design? How to see which file systems are being accessed often? Any suggestions on how to debug further?
Thanks.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-12-2002 12:54 PM
10-12-2002 12:54 PM
Re: IO wait
If I remember correctly, there are some patches to correct sar reporting the wrong IO activity. I'd suggest first using one of the patch bundles available, preferably a newer one.
live free or die
harry
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-12-2002 01:40 PM
10-12-2002 01:40 PM
Re: IO wait
If it does agree, however, then it probably does have something to do with either your database design/layout, or your hardware (disk array) or it's configuation (including LVM config).
Good luck!
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-12-2002 11:47 PM
10-12-2002 11:47 PM
Re: IO wait
From vmstat output in b column there is always number of blocked jobs ranging from 10-14. It is not a patch problem. We have up to date patch.
Thanks.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-13-2002 02:30 PM
10-13-2002 02:30 PM
Re: IO wait
OS and patch level
model of your server
what type of disk(s) are being utilised
What type of connectivity
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-13-2002 03:21 PM
10-13-2002 03:21 PM
Re: IO wait
What are your mount options on the filesystem(s) you are using for the database?
JP
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-14-2002 05:23 AM
10-14-2002 05:23 AM
Re: IO wait
HP-UX 11.00 March 2002 bundles applied
N-4000 4 CPUs < 50% 6 GB RAM
4 Tachyon A5158a HBAs
EMC DS16B (Brocade) FC fabric switches (1GB)
EMC Symmetrix 8730
My Theory-
I think the overall latency to satisfy disk IO requests has been reduced. I suspect that sar shows the serialization of the IO's from multiple systems to the Symmetrix FA occurring at the switch level. I don't think the amount of time in IO wait would be noticeable if the system were busier and had something else to do while the switch merges the traffic into single lanes (6 MB peak switch throughput).
Now I need to prove what's happening (how?) and find a more valid metric, if sar IO wait is no longer a valid indicator of IO performance.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-14-2002 06:44 AM
10-14-2002 06:44 AM
Re: IO wait
Hmm, we run EMC and Brocade switches on lots and lots of HP servers of all classes and OS versions (10.20/11/11i) and we never see wio% > single digits (ie. <10) on all servers even at peak times.
Normally a wio% of <10 is fine, 10-20 means your disk subsystem is having trouble keeping up with i/o requests and >20 means your i/o bound and performance is suffering considerably as a result. If you have wio% of 40-50 then either something is wrong with sar reporting it (does glance confirm or not?) and if sar and glance confirm this 40-50 then your completely i/o bound and should be looking into improving your i/o throughput (EMC cache sizes/config/weighting), more channels etc.
I would certainly be very worried if my wio% was that high.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-14-2002 07:00 AM
10-14-2002 07:00 AM
Re: IO wait
When you are CPU bound, your system is going as fast as it can.
So, I think you guys misunderstand what WAIT IO means...
Processes waiting on I/O spin; which means that when they get a timeslice to run, they check if the I/O has completed, and if not, it idles until the timeslice expires, in the hopes that the I/O will complete before the timeslice ends. This behavior consumes CPU time.
WAIT IO is a measurement of this CPU consumption.
Now, WAIT IO time can be caused by several factors. The most common cause is that the disk array is overloaded, or you have configured it in a non-optimal way - such as putting your logs and dataspaces on a single mirror pair.
So, if you are seeing high WAIT IO (over 10% is high in my opinion), you need to take a good look at your disk array and it's configuration.
You may not have striped over a sufficient number of volumes (not using enough drives), or the disk array may have an internal bottleneck, such as too many systems hitting the same FC port, backplane bottleneck, etc.
Let us know how you make out.
Good luck!