MSA Storage
1752790 Members
6405 Online
108789 Solutions
New Discussion

MSA2324i vDisk Rebuild Problem

 
SOLVED
Go to solution
InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

Thanks for you assistance.

I think I have at least 3 faulty disks as it shows the same error with different serial numbers :

There is a problem with a FRU. (FRU type: disk, enclosure: 1, device ID: 3, vendor: HP , product ID: EG0600FBVFP , SN: KWxxxxx, version: HPDC, related event serial number: A66399, related event code: 55)

This error gets generated when the MSA does the disk scrub i think. However it still shows a green tick on the disk in the enclousure view. 

I have ordered 5 brand new disks to be added tomorrow to the enclosure.

Would it be best to add the new disks, configure as dedicated spares, then use the trust command do you think?

Can I send the logs to you in private message to have a look?

Thanks

 

 

Re: MSA2324i vDisk Rebuild Problem

You need to clearly tell me below details,

1> How many total drives present in your system initially before you tried metadata clear?

2> How many total drives configured as spare before you tried metadata clear?

3> Ofcourse multiple drives having issue otherwise Vdisk can't go into offline state

4> Location of the drive which you have cleared metadata?

5> Do you have sub-vdisk details means which drive paired with which drive to make one RAID set, like that I need details of the 4 RAID set

You can send me the logs in private as well. I can give it a try.

If you have valid support with HPE then it's always good to log a support case officially.

 

Hope this helps!
Regards
Subhajit

I am an HPE employee

If you feel this was helpful please click the KUDOS! thumb below!

**************************************************************************

 

 


I work for HPE
Accept or Kudo
InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

Hi,

Answers below :

1) 15 drives present in enclosure

2) Zero drives Configured as spare

3) .

4) Slots 4, 7, 9 and 13

5) 

Disk1-1 Raid1
Disk1-2 Raid1
Disk1-3 Raid2
Disk1-5 Raid2
Disk1-8 Raid3
Disk1-10 Raid3
Disk1-11 Raid4
Disk1-13 Spare

I dont know how to attach logs, I cant see any upload button in the DM. 

Re: MSA2324i vDisk Rebuild Problem

So I understand RAID sets as below right now,

RAID set 1-> 1,2

RAID set 2-> 3,5

RAID set 3-> 8,10

RAID set 4 -> 11, ?

So you say 4, 7 and 9 were not part of any of the above RAID sets, only 13 was part of them. This can't be possible because drives can only go to LEFTOVER state when they were part of some vdisk and for some reason they marked as LEFTOVER by controllers.

Do you remember how was the RAID sets before drives went to LEFTOVER state?

4 drives metadata clear always risky for data recovery.

You can be extremely lucky if none of 2 drives out of 4,7,9 and 13 part of same RAID set. Then we have good chance to recover this vdisk.

Without logs it's difficult but if you can tell me condition of RAID sets when everything was fine, then which drive join which RAID set after you cleared metadata that may help us. For example, lets say RAID 2 set earlier drive 3 and 4 was in pair but later drive at location went to LEFTOVER state. Then drive at location 5 you have configured as spare and it joins drive 3 as partner. Sameway other sub-vdisks also.

 

Hope this helps!
Regards
Subhajit

I am an HPE employee

If you feel this was helpful please click the KUDOS! thumb below!

**************************************************************************


I work for HPE
Accept or Kudo
InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

Hi,

THanks, for taking time to look at this - really appreciate it.

Luckily I took a screen shot when everything was ok as follows :

vDisk1

RAID set 1-> 1,2

RAID set 2-> 3,5

RAID set 3-> 8,10

RAID set 4 -> 11, 7

vDisk2

RAID set 1-> 6,14

RAID set 2-> 12,15

Disk 13 - must have been a global spare at that particular time.

Not sure about Disk 4,9 though.

 

Sorry I should have mentioned I have a 2nd vDisk Volume.

Does that help?

Solution

Re: MSA2324i vDisk Rebuild Problem

So it's possible that 4, 9 and 13 went to LEFTOVER state long back.

Recently 7 went to LEFTOVER state

Going with the fact and information that you have provided you can try below and 13 should participate in rebuilding the vdisk1

trust enable

trust vdisk vDisk1_R10

show vdisks vDisk1_R10

 

Note: You can still log a HPE Support call and take official help before you try the above.

 

Hope this helps!
Regards
Subhajit

I am an HPE employee

If you feel this was helpful please click the KUDOS! thumb below!

**************************************************************************


I work for HPE
Accept or Kudo
InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

Hi,

Unfortunately we only have hardware support only this MSA so I am unable to log a call with HP.

I am waiting for the 5 brand new disks to be installed first, before I use the trust command.

I think it is best to make disk 13 back to "avail" and configure the new disks as dedicated spares, so the volume would rebuild?

What do you think?

Re: MSA2324i vDisk Rebuild Problem

I don't know if drive 13 having medium errors or any kind of hardware errors or not. If so then you should wait for new drives. Otherwise you can use any of 4,7,9 and 13 for this Vdisk1 data recovery if none of them having any hardware errors.

 

Hope this helps!
Regards
Subhajit

I am an HPE employee

If you feel this was helpful please click the KUDOS! thumb below!

***********************************************************************


I work for HPE
Accept or Kudo
InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

Thanks,

Will let you know how it goes.

InTech2000
Advisor

Re: MSA2324i vDisk Rebuild Problem

We added the 5 disks on friday bays 16-20, and we had some issues with the controllers going offline.

Manage to restart the controller, and configure the 4 out of the 5 as dedicated spares.

We did the "trust" command and the volume came back online and it did a reconstruct momentarily. and it went offline again.. It now showed RAID set 4 -> 16 only - previously is was disk 11.

We did the trust" command and the volume came back online and it did a media scrub.  

It all shows online now and configured as follows :

vDisk1

RAID set 1-> 1,2

RAID set 2-> 3,5

RAID set 3-> 8,10

RAID set 4 -> 11, 16

vDisk2

RAID set 1-> 6,14

RAID set 2-> 12,15

According to the previous logs, they have been issues wth Disk 8 and 11, but at moment,  its showing all green, ticks  vdisk is online and health status "ok". 

Whats the best next steps, I am worried disk 8 and 11 may fail together? Is they any commands I can run to do a check on the disks?

Also, is it better to configure as dedicated or dynamic spare?

Thanks.