<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: linux backup of huge data in Operating System - Linux</title>
    <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193972#M9951</link>
    <description>Hello,&lt;BR /&gt;&lt;BR /&gt;the first problem you have to solve is if a backup that encompasses many files that have been stored over an extended period of time (independent if it is days or just a "few" hours) will do you any good at all. If there is anything in these files that does a coordinated update of more than one of these files you can perform a backup like you do it right now to /dev/null. This will be much faster, cheaper and equally usefull. (Sorry for being so direct but you really need to think about this!)&lt;BR /&gt;&lt;BR /&gt;Now assuming all of these files are totally independent, what most probably happens is that other processes access the filesystem&lt;BR /&gt;while you do your backup and you just do not get enough bandwidth from the disks to your tapedrive to keep the tape streaming. A simple audible check of the drive during the backup should be able to confirm this. &lt;BR /&gt;Things you can do:&lt;BR /&gt;Put the data on a mirror set. Break the set at backuptime, then remount one half readonly and use this to feed the tape. (This will also solve the first problem mentioned. You still might need to quiesce any open databases you have on the volumes before breaking the mirror). It also might help to put the tape on a separate SCSI controller to make sure the bus is not busy with other requests while feeding the tape.&lt;BR /&gt;Given the requirements you have, you should seriously consider an enterprise backup solution. Note: If you do go for something like Veritas or Legato plan on deicated servers and storage.&lt;BR /&gt;&lt;BR /&gt;Greetings, Martin</description>
    <pubDate>Tue, 17 Feb 2004 22:45:15 GMT</pubDate>
    <dc:creator>Martin P.J. Zinser</dc:creator>
    <dc:date>2004-02-17T22:45:15Z</dc:date>
    <item>
      <title>linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193964#M9943</link>
      <description>Hi,&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;I have got 400GB data on a RH Linux Server. Around 350GB data is in a single dir but consisting of hundreds of database files. Now we need to schedule backup. The box has single tape drive of 100/200 GB Ultrium. When I do a manual backup it takes two days and three Ultrium tapes to complete and while the backup continues the system becomes very slow also.&lt;BR /&gt;&lt;BR /&gt; &lt;BR /&gt;&lt;BR /&gt;Can anyone please let me know what are the good options to implement the backup strategy for the box. Now I have to make a backup policy and make it on a cron job and hand over it to the Operations group to carry on with the backup. I am confused how it can be arranged with a backup that needs multi tapes. The It is an IBM intel based server. Is there any other backup tool other than tar to use with Redhat. I remember reading about Amanda. Does anyone have got good experience with it with huge data.&lt;BR /&gt;&lt;BR /&gt; &lt;BR /&gt;&lt;BR /&gt;Thanks&lt;BR /&gt;</description>
      <pubDate>Tue, 17 Feb 2004 07:55:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193964#M9943</guid>
      <dc:creator>Rasheed Tamton</dc:creator>
      <dc:date>2004-02-17T07:55:44Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193965#M9944</link>
      <description>If you dont have an enterprise backup solution like legato or veritas it will take a long time to back up this data, as you know. &lt;BR /&gt;&lt;BR /&gt;On a couple of my RH9 systems I use a product called 'star'. It is easy to download and install. It runs a bit faster than traditional 'tar', but it is still going to take a long time.&lt;BR /&gt;&lt;BR /&gt;Another option, if you have lots of capacity, is to copy the data to another silo and back it up there. Taht way, time is of no concern.&lt;BR /&gt;&lt;BR /&gt;If this is an ongoing backup each day or week. You need to look into a single license for veritas. I backup 300 GB in about 3.5 hours. Also, ArcServe is available but not real reliable.&lt;BR /&gt;&lt;BR /&gt;Also, try a product called Amanda. I believe it is part of the install of RedHat version9.</description>
      <pubDate>Tue, 17 Feb 2004 08:06:37 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193965#M9944</guid>
      <dc:creator>Nobody's Hero</dc:creator>
      <dc:date>2004-02-17T08:06:37Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193966#M9945</link>
      <description>The only thing I can think of is an SDLT or LTO tape drive. You will get all your data on one tape with hardware compression and they are fast.  We sometimes restore from SDLT tape rather than use "cp" because it is faster.</description>
      <pubDate>Tue, 17 Feb 2004 08:44:58 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193966#M9945</guid>
      <dc:creator>Mark Grant</dc:creator>
      <dc:date>2004-02-17T08:44:58Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193967#M9946</link>
      <description>Can't you do incremental backups ? SAy incremental everyday and full backup on week ends ?&lt;BR /&gt;&lt;BR /&gt;Amanda is said to be slightly hard to use, star ( &lt;A href="http://www.fokus.gmd.de/research/cc/glone/employees/joerg.schilling/private/star.html" target="_blank"&gt;http://www.fokus.gmd.de/research/cc/glone/employees/joerg.schilling/private/star.html&lt;/A&gt; ) works great, I like much mondo ( &lt;A href="http://www.microwerks.net/~hugo/" target="_blank"&gt;http://www.microwerks.net/~hugo/&lt;/A&gt; ) too...&lt;BR /&gt;&lt;BR /&gt;J</description>
      <pubDate>Tue, 17 Feb 2004 08:59:57 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193967#M9946</guid>
      <dc:creator>Jerome Henry</dc:creator>
      <dc:date>2004-02-17T08:59:57Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193968#M9947</link>
      <description>Hi Rasheed,&lt;BR /&gt;&lt;BR /&gt;I assuming you are using Ultrium 215, the slower module in LTO-1 family. Even without compression, you should be able to reach 27GB/hr. That means, for a full backup of 400GB data, the time should be less than a day. With H/W compression enabled, it should be faster. We can see the compression is enabled since you are not using 4 100/200G tapes for 400G data.&lt;BR /&gt;&lt;BR /&gt;Ultrium drive backup performance issue might also caused by system resources. Recommanded 1G Hz CPU and 1 GB memory, it also depends on how much work load do you have on the server. You can download HP Library &amp;amp; Tapetools from &lt;A href="http://www.hp.com/support/tapetools," target="_blank"&gt;http://www.hp.com/support/tapetools,&lt;/A&gt; and run a sys perf to analysis the server throughput. Maybe the server is the bottle neck.&lt;BR /&gt;&lt;BR /&gt;Besides that, changing to use incremental backup is a good way to reduce the amount of data being backup.&lt;BR /&gt;&lt;BR /&gt;Brice</description>
      <pubDate>Tue, 17 Feb 2004 14:31:42 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193968#M9947</guid>
      <dc:creator>Brice_3</dc:creator>
      <dc:date>2004-02-17T14:31:42Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193969#M9948</link>
      <description>It takes way too much time time for your backup, you have a performance problem.&lt;BR /&gt;You should hve good performance with dump/restore ( standard on redhat ). &lt;BR /&gt;I used other soft like tar, gtar  but none of them could feed an utltrium as fast as the HW can take.&lt;BR /&gt;</description>
      <pubDate>Tue, 17 Feb 2004 14:57:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193969#M9948</guid>
      <dc:creator>Olivier Drouin</dc:creator>
      <dc:date>2004-02-17T14:57:14Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193970#M9949</link>
      <description>I bet you have millions of little files.&lt;BR /&gt;&lt;BR /&gt;Millions of little files can cause something even like Legato and a fast tape subsystem to take days to restore.&lt;BR /&gt;&lt;BR /&gt;Though without knowing more about this particular environment this is just a guess.&lt;BR /&gt;&lt;BR /&gt;I recommend that you find the bottleneck and then look to widening it.&lt;BR /&gt;&lt;BR /&gt;Like I speculated above, if you have lots of little files then the bottleneck is probably the OS doing open/close on the little files. &lt;BR /&gt;&lt;BR /&gt;To test this, try backup/restoring a single file of 50G or so.  Just dd something from /dev/urandom.  Make it from /dev/urandom so any compression done in the tape drive won't skew your results.  Then go and backup and restore the thing.  It should go quite fast.&lt;BR /&gt;&lt;BR /&gt;Before discussing a solution you should find  the bottleneck.&lt;BR /&gt;</description>
      <pubDate>Tue, 17 Feb 2004 17:01:24 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193970#M9949</guid>
      <dc:creator>Mark Travis</dc:creator>
      <dc:date>2004-02-17T17:01:24Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193971#M9950</link>
      <description>Oops.  I didn't read closely enough.  If all you have are hundreds of files then filesystem open/close shouldn't cause a bottleneck.&lt;BR /&gt;&lt;BR /&gt;But anyway, advice about finding the bottleneck holds.  You need to find the bottleneck.</description>
      <pubDate>Tue, 17 Feb 2004 17:04:03 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193971#M9950</guid>
      <dc:creator>Mark Travis</dc:creator>
      <dc:date>2004-02-17T17:04:03Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193972#M9951</link>
      <description>Hello,&lt;BR /&gt;&lt;BR /&gt;the first problem you have to solve is if a backup that encompasses many files that have been stored over an extended period of time (independent if it is days or just a "few" hours) will do you any good at all. If there is anything in these files that does a coordinated update of more than one of these files you can perform a backup like you do it right now to /dev/null. This will be much faster, cheaper and equally usefull. (Sorry for being so direct but you really need to think about this!)&lt;BR /&gt;&lt;BR /&gt;Now assuming all of these files are totally independent, what most probably happens is that other processes access the filesystem&lt;BR /&gt;while you do your backup and you just do not get enough bandwidth from the disks to your tapedrive to keep the tape streaming. A simple audible check of the drive during the backup should be able to confirm this. &lt;BR /&gt;Things you can do:&lt;BR /&gt;Put the data on a mirror set. Break the set at backuptime, then remount one half readonly and use this to feed the tape. (This will also solve the first problem mentioned. You still might need to quiesce any open databases you have on the volumes before breaking the mirror). It also might help to put the tape on a separate SCSI controller to make sure the bus is not busy with other requests while feeding the tape.&lt;BR /&gt;Given the requirements you have, you should seriously consider an enterprise backup solution. Note: If you do go for something like Veritas or Legato plan on deicated servers and storage.&lt;BR /&gt;&lt;BR /&gt;Greetings, Martin</description>
      <pubDate>Tue, 17 Feb 2004 22:45:15 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193972#M9951</guid>
      <dc:creator>Martin P.J. Zinser</dc:creator>
      <dc:date>2004-02-17T22:45:15Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193973#M9952</link>
      <description>There is a new Ultrium format generation 2 tapes which do handle up to 400 GB of data. That can do the job for you.  HP sells them and some of them are certified for Linux.&lt;BR /&gt;&lt;BR /&gt;If funding is available, you might want to investigate one of HP's tape libraries. These can accommodate up to 4 drives and many tapes.&lt;BR /&gt;&lt;BR /&gt;We have implemented such a solution at our shop.&lt;BR /&gt;&lt;BR /&gt;SEP</description>
      <pubDate>Tue, 17 Feb 2004 22:59:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193973#M9952</guid>
      <dc:creator>Steven E. Protter</dc:creator>
      <dc:date>2004-02-17T22:59:01Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193974#M9953</link>
      <description>Thanks to everyone who commented on the subject. It is an IBM eServer X series 250. The database is mySQL. I just need a filesystem backup to the tape. Tape drive is (Vendor: HP Model: Ultrium 1-SCSI   Rev: N15G). &lt;BR /&gt;&lt;BR /&gt;I use 200GB Ultrium which can have a hardware compressed files of 200GB. But most of the files are mySQL database files I do not get the HW compression from the tape. So I normally use 3 tapes to get the 400GB data.&lt;BR /&gt;&lt;BR /&gt;We have got Legato and BrightStore with other systems but with this particular box I am not sure we will be going for those. The data is from a legacy mainframe that do not change much often. Normally used for reports - not updates. I need to do the backup once per week only.&lt;BR /&gt;It has dual PIII processors as below&lt;BR /&gt;model name      : Pentium III (Cascades)&lt;BR /&gt;cpu MHz         : 699.199&lt;BR /&gt;cache size      : 1024 KB&lt;BR /&gt;Phys Mem : 16GB&lt;BR /&gt;&lt;BR /&gt;I do not have enough free space to move it to another file system or to make mirror.&lt;BR /&gt;&lt;BR /&gt;Pls advise.&lt;BR /&gt;Thanks&lt;BR /&gt;Rasheed.&lt;BR /&gt;</description>
      <pubDate>Wed, 18 Feb 2004 06:54:01 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193974#M9953</guid>
      <dc:creator>Rasheed Tamton</dc:creator>
      <dc:date>2004-02-18T06:54:01Z</dc:date>
    </item>
    <item>
      <title>Re: linux backup of huge data</title>
      <link>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193975#M9954</link>
      <description>Only an idea for more performance with tar.&lt;BR /&gt;Would it be possible to remount the filesystem with "noatime" for backup? &lt;BR /&gt;With the default atime each read to a file (and even the backup by tar) updates the access time entry for the file.&lt;BR /&gt;Sometimes software compression is faster. &lt;BR /&gt;&lt;BR /&gt;Have you tried the z option with tar? (but better don't combine hardware AND software compression).&lt;BR /&gt;&lt;BR /&gt;dump/restore is an idea, but my last experience is that the file system has to be in a quiet state (remounted read only or unmounted). In the other case the the backup seems to be okay, but there are "block expected/different block found errors while restore.</description>
      <pubDate>Mon, 23 Feb 2004 10:10:14 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-linux/linux-backup-of-huge-data/m-p/3193975#M9954</guid>
      <dc:creator>Thilo Knoch</dc:creator>
      <dc:date>2004-02-23T10:10:14Z</dc:date>
    </item>
  </channel>
</rss>

