- Community Home
- >
- Storage
- >
- Entry Storage Systems
- >
- Disk Enclosures
- >
- MSA50s failed with high I/O load
Disk Enclosures
1748245
Members
3774
Online
108760
Solutions
Forums
Categories
Company
Local Language
back
Forums
Discussions
Forums
- Data Protection and Retention
- Entry Storage Systems
- Legacy
- Midrange and Enterprise Storage
- Storage Networking
- HPE Nimble Storage
Discussions
Discussions
Discussions
Forums
Discussions
back
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
- BladeSystem Infrastructure and Application Solutions
- Appliance Servers
- Alpha Servers
- BackOffice Products
- Internet Products
- HPE 9000 and HPE e3000 Servers
- Networking
- Netservers
- Secure OS Software for Linux
- Server Management (Insight Manager 7)
- Windows Server 2003
- Operating System - Tru64 Unix
- ProLiant Deployment and Provisioning
- Linux-Based Community / Regional
- Microsoft System Center Integration
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Discussion Boards
Blogs
Information
Community
Resources
Community Language
Language
Forums
Blogs
Topic Options
- Subscribe to RSS Feed
- Mark Topic as New
- Mark Topic as Read
- Float this Topic for Current User
- Bookmark
- Subscribe
- Printer Friendly Page
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-06-2007 05:47 AM
08-06-2007 05:47 AM
MSA50s failed with high I/O load
Had a very strange failure with some MSA50s this weekend.
Here's the configuration:
ML530G2
2x P600
3x MSA50 w/ 10 - 72GB 10k SAS SFF
Changed the arrays from 3 10-disk RAID 10 to 1 20-disk RAID10 left the 1 10-disk.
Started a 40GB file copy from another server (10-disk RAID10 w/ 15k SAS drives) over 1Gb network to the 20-disk RAID 10. During that transfer, which only took a few minutes, two disks in the 20-disk array on the MSA50s tripped offline. One of them with failure code 7, the other with code 32.
I reseated those drives and they came back online with no problem and rebuilt from their mirrored pair.
Then I started restoring a database from that 40GB file to those arrays on the MSA50s. At that point, the drives started lighting up like a Christmas tree. I lost ten drives in the 20-disk RAID10 (one of each of the mirrored pairs), and on the 10-disk array, I lost three.
All of these failures were either failure code 7, 20 or 32. There are 4 SCSI bus faults on all drives in the arrays and on the drives that show failed, under "Other Failures" there is a count of 3.
All of the disks are model# DG072A8B54 and are running an older version of firmare, HPD4.
Unfortunately, HP does not show what updates were made in the subsequent 3 firmware releases for these drives. Nor have I ever received an alert concerning these SAS SFF drives requiring firmware updates.
The thing that shocks me the most is that this server had an I/O load on it before the reconfigure and the tasks I performed this weekend. Why did this setup crap out on me like this now?
I've created a case with HP and set in a few ADU reports from different times during these failures. Hopefully they can come up with something other than, "You need to update your firmware." I need to know why this failure happened and assurance that firmware HPD7 will fix it for good...
Here's the configuration:
ML530G2
2x P600
3x MSA50 w/ 10 - 72GB 10k SAS SFF
Changed the arrays from 3 10-disk RAID 10 to 1 20-disk RAID10 left the 1 10-disk.
Started a 40GB file copy from another server (10-disk RAID10 w/ 15k SAS drives) over 1Gb network to the 20-disk RAID 10. During that transfer, which only took a few minutes, two disks in the 20-disk array on the MSA50s tripped offline. One of them with failure code 7, the other with code 32.
I reseated those drives and they came back online with no problem and rebuilt from their mirrored pair.
Then I started restoring a database from that 40GB file to those arrays on the MSA50s. At that point, the drives started lighting up like a Christmas tree. I lost ten drives in the 20-disk RAID10 (one of each of the mirrored pairs), and on the 10-disk array, I lost three.
All of these failures were either failure code 7, 20 or 32. There are 4 SCSI bus faults on all drives in the arrays and on the drives that show failed, under "Other Failures" there is a count of 3.
All of the disks are model# DG072A8B54 and are running an older version of firmare, HPD4.
Unfortunately, HP does not show what updates were made in the subsequent 3 firmware releases for these drives. Nor have I ever received an alert concerning these SAS SFF drives requiring firmware updates.
The thing that shocks me the most is that this server had an I/O load on it before the reconfigure and the tasks I performed this weekend. Why did this setup crap out on me like this now?
I've created a case with HP and set in a few ADU reports from different times during these failures. Hopefully they can come up with something other than, "You need to update your firmware." I need to know why this failure happened and assurance that firmware HPD7 will fix it for good...
The opinions expressed above are the personal opinions of the authors, not of Hewlett Packard Enterprise. By using this site, you accept the Terms of Use and Rules of Participation.
News and Events
Support
© Copyright 2024 Hewlett Packard Enterprise Development LP