<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: SLES 10 SP1, DRBD, and rsync in Operating System - Linux</title>
    <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098205#M30663</link>
    <description>we used to have a problem with our attached storage going read-only quite often.  I found an old bug report at the time that pointed to SCSI timeouts.   Might not be the same problem, but might help.  &lt;A href="http://www.redhatmagazine.com/2006/12/15/tips_tricks/" target="_blank"&gt;http://www.redhatmagazine.com/2006/12/15/tips_tricks/&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;It is about half way down titled &lt;BR /&gt;Why does the ext3 filesystems on my Storage Area Network (SAN) repeatedly become read-only?</description>
    <pubDate>Tue, 06 Nov 2007 18:28:14 GMT</pubDate>
    <dc:creator>Justin_99</dc:creator>
    <dc:date>2007-11-06T18:28:14Z</dc:date>
    <item>
      <title>SLES 10 SP1, DRBD, and rsync</title>
      <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098203#M30661</link>
      <description>Weird situation that I haven't been able to figure out yet. Thought someone on here may have run into it.&lt;BR /&gt;&lt;BR /&gt;I have two Heartbeat/DRBD clusters configured identically (other than names, of course). I was trying to sync the data between the current non-clustered nodes and the new clusters. On one cluster, this appears to work perfectly using rsync. On the other, the file system becomes read-only at seemingly random times. I've been running a tar piped via ssh to do a blind copy of the data to this cluster and that's been running smoothly for a while now, but any attempt to use rsync results in the file system going read-only. Not ideal for the final cut over.&lt;BR /&gt;&lt;BR /&gt;I didn't find anything useful with a Google search. Any ideas?</description>
      <pubDate>Tue, 06 Nov 2007 15:49:18 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098203#M30661</guid>
      <dc:creator>Jeff_Traigle</dc:creator>
      <dc:date>2007-11-06T15:49:18Z</dc:date>
    </item>
    <item>
      <title>Re: SLES 10 SP1, DRBD, and rsync</title>
      <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098204#M30662</link>
      <description>Shalom,&lt;BR /&gt;&lt;BR /&gt;Looks to me like a network inconsistency. A filesystem in Linux goes read only when there is a problem.&lt;BR /&gt;&lt;BR /&gt;Here is what I'd check on:&lt;BR /&gt;1) dmesg, look for a problem related to the disk that underlies the filesystem.&lt;BR /&gt;2) fsck the filesystem itself after umounting it. Same as hp-ux, no fsck with the filesystem hot.&lt;BR /&gt;3) Consider building a new ram disk (mkinitrd) on the effected system.&lt;BR /&gt;&lt;BR /&gt;There is evidence, I'd like to see it to help guide the diagnosis.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;SEP</description>
      <pubDate>Tue, 06 Nov 2007 16:22:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098204#M30662</guid>
      <dc:creator>Steven E. Protter</dc:creator>
      <dc:date>2007-11-06T16:22:01Z</dc:date>
    </item>
    <item>
      <title>Re: SLES 10 SP1, DRBD, and rsync</title>
      <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098205#M30663</link>
      <description>we used to have a problem with our attached storage going read-only quite often.  I found an old bug report at the time that pointed to SCSI timeouts.   Might not be the same problem, but might help.  &lt;A href="http://www.redhatmagazine.com/2006/12/15/tips_tricks/" target="_blank"&gt;http://www.redhatmagazine.com/2006/12/15/tips_tricks/&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;It is about half way down titled &lt;BR /&gt;Why does the ext3 filesystems on my Storage Area Network (SAN) repeatedly become read-only?</description>
      <pubDate>Tue, 06 Nov 2007 18:28:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098205#M30663</guid>
      <dc:creator>Justin_99</dc:creator>
      <dc:date>2007-11-06T18:28:14Z</dc:date>
    </item>
    <item>
      <title>Re: SLES 10 SP1, DRBD, and rsync</title>
      <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098206#M30664</link>
      <description>Nothing in dmesg. While no fsck is required to remount it read-write after it does this, fsck did detect a problem:&lt;BR /&gt;&lt;BR /&gt;host1:~ # fsck /dev/drbd0&lt;BR /&gt;fsck 1.38 (30-Jun-2005)&lt;BR /&gt;e2fsck 1.38 (30-Jun-2005)&lt;BR /&gt;/dev/drbd0: recovering journal&lt;BR /&gt;The filesystem size (according to the superblock) is 53739520 blocks&lt;BR /&gt;The physical size of the device is 53287936 blocks&lt;BR /&gt;Either the superblock or the partition table is likely to be corrupt!&lt;BR /&gt;&lt;BR /&gt;This seems to correlate with some syslog messages I found this morning:&lt;BR /&gt;&lt;BR /&gt;Nov  6 17:09:46 fshare1 kernel: attempt to access beyond end of device&lt;BR /&gt;Nov  6 17:09:46 fshare1 kernel: drbd0: rw=0, want=426770448, limit=426303488&lt;BR /&gt;Nov  6 17:09:46 fshare1 kernel: EXT3-fs error (device drbd0): read_inode_bitmap: Cannot read inode bitmap - block_group = 1628, inode_bitmap = 53346305&lt;BR /&gt;Nov  6 17:09:46 fshare1 kernel: Aborting journal on device drbd0.&lt;BR /&gt;Nov  6 17:09:46 fshare1 kernel: EXT3-fs error (device drbd0) in ext3_ordered_writepage: IO failure&lt;BR /&gt;Nov  6 17:09:47 fshare1 kernel: ext3_abort called.&lt;BR /&gt;Nov  6 17:09:47 fshare1 kernel: EXT3-fs error (device drbd0): ext3_journal_start_sb: Detected aborted journal&lt;BR /&gt;Nov  6 17:09:47 fshare1 kernel: Remounting filesystem read-only&lt;BR /&gt;Nov  6 17:09:50 fshare1 kernel: EXT3-fs error (device drbd0) in ext3_new_inode: IO failure&lt;BR /&gt;Nov  6 17:09:50 fshare1 kernel: EXT3-fs error (device drbd0) in ext3_create: IO failure&lt;BR /&gt;Nov  6 17:10:09 fshare1 kernel: __journal_remove_journal_head: freeing b_committed_data&lt;BR /&gt;&lt;BR /&gt;All of which led me to the discovery this morning that the available space for the RAID-5 LUN on the internal arrays don't match (203.4GB on one and 205.0GB on the other). First rule of clusters is make everything identical. :) So I'm reformatting the LVs to the lower capacity so they match and will test that. I'm pretty sure that should fix the problem though.</description>
      <pubDate>Wed, 07 Nov 2007 10:29:13 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098206#M30664</guid>
      <dc:creator>Jeff_Traigle</dc:creator>
      <dc:date>2007-11-07T10:29:13Z</dc:date>
    </item>
    <item>
      <title>Re: SLES 10 SP1, DRBD, and rsync</title>
      <link>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098207#M30665</link>
      <description>Well, it looked promising, but last night I continued having the problem (although it doesn't seem to happen as often now... I was able to copy all the data after a 3 attempts). So, I have the following setup:&lt;BR /&gt;&lt;BR /&gt;1. The mismatched physical size of the LUNs (203.4GB on one and 205.0GB on the other.&lt;BR /&gt;2. The matching LVs defined using these LUNs (both 203.4GB).&lt;BR /&gt;3. DRBD configured to use these matching LVs.&lt;BR /&gt;&lt;BR /&gt;I got the same "attempt to access beyond end of device" error registered in syslog a few hours ago. This time, fsck found no problems, however:&lt;BR /&gt;&lt;BR /&gt;hostname1:~ # fsck /dev/mapper/vg01-data&lt;BR /&gt;fsck 1.38 (30-Jun-2005)&lt;BR /&gt;e2fsck 1.38 (30-Jun-2005)&lt;BR /&gt;/dev/mapper/vg01-data: recovering journal&lt;BR /&gt;/dev/mapper/vg01-data contains a file system with errors, check forced.&lt;BR /&gt;Pass 1: Checking inodes, blocks, and sizes&lt;BR /&gt;Pass 2: Checking directory structure&lt;BR /&gt;Pass 3: Checking directory connectivity&lt;BR /&gt;Pass 4: Checking reference counts&lt;BR /&gt;Pass 5: Checking group summary information&lt;BR /&gt;/dev/mapper/vg01-data: 175608/26673152 files (1.7% non-contiguous), 36824434/53320704 blocks&lt;BR /&gt;&lt;BR /&gt;So I'm confused. Based on the message, it seems like DRBD is trying to write to the space on the LUN on the primary system that isn't being used in the underlying LV?</description>
      <pubDate>Fri, 09 Nov 2007 08:55:02 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/sles-10-sp1-drbd-and-rsync/m-p/4098207#M30665</guid>
      <dc:creator>Jeff_Traigle</dc:creator>
      <dc:date>2007-11-09T08:55:02Z</dc:date>
    </item>
  </channel>
</rss>

