<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic MD Raid system VG crashes with more than 1 md device in Operating System - Linux</title>
    <link>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447913#M37195</link>
    <description>&lt;!--!*#--&gt;NOTE: This is a standard Redhat 5.2 build on a DL580&lt;BR /&gt;&lt;BR /&gt;I started by creating a "system" LVM VG (containing root and swap) that contains a number of MD raid devices, as per this thread:&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://forums11.itrc.hp.com/service/forums/questionanswer.do?threadId=1349124" target="_blank"&gt;http://forums11.itrc.hp.com/service/forums/questionanswer.do?threadId=1349124&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;But what I've found is that after the VG is extended to contain the extra md devices, the system crashes on reboot with this error:&lt;BR /&gt;&lt;BR /&gt;Red Hat nash version 5.1.19.6 starting&lt;BR /&gt;  Reading all physical volumes. This may take a while...&lt;BR /&gt;  Couldn't find device with uuid 'xXG4FE-rq9K-IZvZ-uxQx-u2fv-FgWq-0AeI2T'.&lt;BR /&gt;  Couldn't find all physical volumes for the volume group system.&lt;BR /&gt;&lt;BR /&gt;See the attached screenshot for the full set of messages.  It's just fine when it's just the one md device (md1).  &lt;BR /&gt;&lt;BR /&gt;The UUID it's complaining about is the UUID for the next md device in the list, md2:&lt;BR /&gt;&lt;BR /&gt;  --- Physical volumes ---&lt;BR /&gt;  PV Name               /dev/md1&lt;BR /&gt;  PV UUID               jPe1pE-MAzL-dn0g-N6gd-PX0s-Y7wO-dLAIkf&lt;BR /&gt;  PV Status             allocatable&lt;BR /&gt;  Total PE / Free PE    4364 / 2252&lt;BR /&gt;&lt;BR /&gt;  PV Name               /dev/md2&lt;BR /&gt;  PV UUID               xXG4FE-rq9K-IZvZ-uxQx-u2fv-FgWq-0AeI2T&lt;BR /&gt;  PV Status             allocatable&lt;BR /&gt;  Total PE / Free PE    4374 / 4374&lt;BR /&gt;&lt;BR /&gt;I had a look at the nash init script in the initrd image and sure enough it was only starting the one md device:&lt;BR /&gt;&lt;BR /&gt;echo Scanning and configuring dmraid supported devices&lt;BR /&gt;raidautorun /dev/md1&lt;BR /&gt;echo Scanning logical volumes&lt;BR /&gt;lvm vgscan --ignorelockingfailure&lt;BR /&gt;&lt;BR /&gt;So I remade initrd, forcing it to probe for raid volumes and that on the surface appeared to improve things.&lt;BR /&gt;&lt;BR /&gt;# mkinitrd -f --force-raid-probe /boot/initrd-2.6.18-92.el5.img 2.6.18-92.el5&lt;BR /&gt;&lt;BR /&gt;The new initrd image now contains an init nash script updated with all the md devices:&lt;BR /&gt;&lt;BR /&gt;echo Scanning and configuring dmraid supported devices&lt;BR /&gt;raidautorun /dev/md1&lt;BR /&gt;raidautorun /dev/md2&lt;BR /&gt;raidautorun /dev/md3&lt;BR /&gt;raidautorun /dev/md4&lt;BR /&gt;raidautorun /dev/md5&lt;BR /&gt;raidautorun /dev/md6&lt;BR /&gt;raidautorun /dev/md7&lt;BR /&gt;raidautorun /dev/md8&lt;BR /&gt;echo Scanning logical volumes&lt;BR /&gt;lvm vgscan --ignorelockingfailure&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;I thought I had it licked at that point, but the system still crashes in this configuration with the same error.  &lt;BR /&gt;&lt;BR /&gt;What am I missing?  Could this be a bug? If raidautorun successfully assembles the md devices then the lvm vgscan should pick them up surely. &lt;BR /&gt;&lt;BR /&gt;There's nothing wrong with the VG itself. I can boot using a SystemRescueCD iso and assemble the system VG with out any issue:&lt;BR /&gt;&lt;BR /&gt;# mdadm -Esb &amp;gt; /etc/mdadm.conf&lt;BR /&gt;# mdadm --assemble --scan&lt;BR /&gt;# vgscan&lt;BR /&gt;# vgchange -a y system&lt;BR /&gt;# vgdisplay -v&lt;BR /&gt;&lt;BR /&gt;The "root" LV is mountable at that point and everything looks good. vgreducing the VG at this point to just md1 gets me a bootable system again.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;&lt;BR /&gt;John&lt;BR /&gt;</description>
    <pubDate>Fri, 26 Jun 2009 11:36:48 GMT</pubDate>
    <dc:creator>John McNulty_2</dc:creator>
    <dc:date>2009-06-26T11:36:48Z</dc:date>
    <item>
      <title>MD Raid system VG crashes with more than 1 md device</title>
      <link>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447913#M37195</link>
      <description>&lt;!--!*#--&gt;NOTE: This is a standard Redhat 5.2 build on a DL580&lt;BR /&gt;&lt;BR /&gt;I started by creating a "system" LVM VG (containing root and swap) that contains a number of MD raid devices, as per this thread:&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://forums11.itrc.hp.com/service/forums/questionanswer.do?threadId=1349124" target="_blank"&gt;http://forums11.itrc.hp.com/service/forums/questionanswer.do?threadId=1349124&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;But what I've found is that after the VG is extended to contain the extra md devices, the system crashes on reboot with this error:&lt;BR /&gt;&lt;BR /&gt;Red Hat nash version 5.1.19.6 starting&lt;BR /&gt;  Reading all physical volumes. This may take a while...&lt;BR /&gt;  Couldn't find device with uuid 'xXG4FE-rq9K-IZvZ-uxQx-u2fv-FgWq-0AeI2T'.&lt;BR /&gt;  Couldn't find all physical volumes for the volume group system.&lt;BR /&gt;&lt;BR /&gt;See the attached screenshot for the full set of messages.  It's just fine when it's just the one md device (md1).  &lt;BR /&gt;&lt;BR /&gt;The UUID it's complaining about is the UUID for the next md device in the list, md2:&lt;BR /&gt;&lt;BR /&gt;  --- Physical volumes ---&lt;BR /&gt;  PV Name               /dev/md1&lt;BR /&gt;  PV UUID               jPe1pE-MAzL-dn0g-N6gd-PX0s-Y7wO-dLAIkf&lt;BR /&gt;  PV Status             allocatable&lt;BR /&gt;  Total PE / Free PE    4364 / 2252&lt;BR /&gt;&lt;BR /&gt;  PV Name               /dev/md2&lt;BR /&gt;  PV UUID               xXG4FE-rq9K-IZvZ-uxQx-u2fv-FgWq-0AeI2T&lt;BR /&gt;  PV Status             allocatable&lt;BR /&gt;  Total PE / Free PE    4374 / 4374&lt;BR /&gt;&lt;BR /&gt;I had a look at the nash init script in the initrd image and sure enough it was only starting the one md device:&lt;BR /&gt;&lt;BR /&gt;echo Scanning and configuring dmraid supported devices&lt;BR /&gt;raidautorun /dev/md1&lt;BR /&gt;echo Scanning logical volumes&lt;BR /&gt;lvm vgscan --ignorelockingfailure&lt;BR /&gt;&lt;BR /&gt;So I remade initrd, forcing it to probe for raid volumes and that on the surface appeared to improve things.&lt;BR /&gt;&lt;BR /&gt;# mkinitrd -f --force-raid-probe /boot/initrd-2.6.18-92.el5.img 2.6.18-92.el5&lt;BR /&gt;&lt;BR /&gt;The new initrd image now contains an init nash script updated with all the md devices:&lt;BR /&gt;&lt;BR /&gt;echo Scanning and configuring dmraid supported devices&lt;BR /&gt;raidautorun /dev/md1&lt;BR /&gt;raidautorun /dev/md2&lt;BR /&gt;raidautorun /dev/md3&lt;BR /&gt;raidautorun /dev/md4&lt;BR /&gt;raidautorun /dev/md5&lt;BR /&gt;raidautorun /dev/md6&lt;BR /&gt;raidautorun /dev/md7&lt;BR /&gt;raidautorun /dev/md8&lt;BR /&gt;echo Scanning logical volumes&lt;BR /&gt;lvm vgscan --ignorelockingfailure&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;I thought I had it licked at that point, but the system still crashes in this configuration with the same error.  &lt;BR /&gt;&lt;BR /&gt;What am I missing?  Could this be a bug? If raidautorun successfully assembles the md devices then the lvm vgscan should pick them up surely. &lt;BR /&gt;&lt;BR /&gt;There's nothing wrong with the VG itself. I can boot using a SystemRescueCD iso and assemble the system VG with out any issue:&lt;BR /&gt;&lt;BR /&gt;# mdadm -Esb &amp;gt; /etc/mdadm.conf&lt;BR /&gt;# mdadm --assemble --scan&lt;BR /&gt;# vgscan&lt;BR /&gt;# vgchange -a y system&lt;BR /&gt;# vgdisplay -v&lt;BR /&gt;&lt;BR /&gt;The "root" LV is mountable at that point and everything looks good. vgreducing the VG at this point to just md1 gets me a bootable system again.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;&lt;BR /&gt;John&lt;BR /&gt;</description>
      <pubDate>Fri, 26 Jun 2009 11:36:48 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447913#M37195</guid>
      <dc:creator>John McNulty_2</dc:creator>
      <dc:date>2009-06-26T11:36:48Z</dc:date>
    </item>
    <item>
      <title>Re: MD Raid system VG crashes with more than 1 md device</title>
      <link>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447914#M37196</link>
      <description>Shalom,&lt;BR /&gt;&lt;BR /&gt;DL580 systems generally ship with a hardware raid system that is a better performance option than using software raid.&lt;BR /&gt;&lt;BR /&gt;I suspect there is a problem with one of the disks.&lt;BR /&gt;&lt;BR /&gt;Take a look at dmesg output and run some dd tests to validate the disks.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;SEP</description>
      <pubDate>Fri, 26 Jun 2009 13:44:43 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447914#M37196</guid>
      <dc:creator>Steven E. Protter</dc:creator>
      <dc:date>2009-06-26T13:44:43Z</dc:date>
    </item>
    <item>
      <title>Re: MD Raid system VG crashes with more than 1 md device</title>
      <link>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447915#M37197</link>
      <description>There is no problem with any of the disks. dmesg shows nothing, and mdstat shows all raid sets fully populated.&lt;BR /&gt;&lt;BR /&gt;# cat /proc/mdstat&lt;BR /&gt;Personalities : [raid1]&lt;BR /&gt;md8 : active raid1 cciss/c0d7p1[0] cciss/c1d7p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md7 : active raid1 cciss/c0d6p1[0] cciss/c1d6p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md6 : active raid1 cciss/c0d5p1[0] cciss/c1d5p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md5 : active raid1 cciss/c0d4p1[0] cciss/c1d4p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md4 : active raid1 cciss/c0d3p1[0] cciss/c1d3p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md3 : active raid1 cciss/c0d2p1[0] cciss/c1d2p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md2 : active raid1 cciss/c0d1p1[0] cciss/c1d1p1[1]&lt;BR /&gt;      143331776 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md0 : active raid1 cciss/c1d0p1[1] cciss/c0d0p1[0]&lt;BR /&gt;      305088 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;md1 : active raid1 cciss/c1d0p2[1] cciss/c0d0p2[0]&lt;BR /&gt;      143026624 blocks [2/2] [UU]&lt;BR /&gt;&lt;BR /&gt;unused devices: &lt;NONE&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;Also about hardware RAID: this system is part of a highly available solution. Database disks will live on an EVA and be mirrored. Dual EVA controllers, dual FC paths via dual FC switches to dual FC HBAs. Dual NICs for 3 networks, and 2 DSL580s to form a dual Oracle 11g RAC cluster. Dual clusters too: one Live and one Data Guard cluster elsewhere on the campus.  &lt;BR /&gt;&lt;BR /&gt;So I'm not going to compromise resilience by putting local data (system disk, etc) on a RAID5 disk attached to a single controller. Especially when this customer has had previous experience of RAID controllers failing in other kit.  Also, RAID10 volumes win over RAID5 in my view anyway.&lt;BR /&gt;&lt;/NONE&gt;</description>
      <pubDate>Fri, 26 Jun 2009 14:14:50 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447915#M37197</guid>
      <dc:creator>John McNulty_2</dc:creator>
      <dc:date>2009-06-26T14:14:50Z</dc:date>
    </item>
    <item>
      <title>Re: MD Raid system VG crashes with more than 1 md device</title>
      <link>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447916#M37198</link>
      <description>&lt;BR /&gt;I've decided to side step this issue completely and just put the O/S volume group into a single md raid device and treat the local data disks separately in their own VG.&lt;BR /&gt;&lt;BR /&gt;Clearly a multi MD system VG is not a configuration many people are familiar with, and I don't want to be "out there" on systems that need to be solid and stable.&lt;BR /&gt;&lt;BR /&gt;Rgds,&lt;BR /&gt;&lt;BR /&gt;John</description>
      <pubDate>Mon, 29 Jun 2009 10:54:20 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/md-raid-system-vg-crashes-with-more-than-1-md-device/m-p/4447916#M37198</guid>
      <dc:creator>John McNulty_2</dc:creator>
      <dc:date>2009-06-29T10:54:20Z</dc:date>
    </item>
  </channel>
</rss>

