<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: millions of files per directory in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985554#M604685</link>
    <description>That's correct. The version 5 directory structure (and associated subsystem code) is much more efficient with directory scans. Note that VxFS (any version) can have essentially unlimited files. What you do to manipulate the the structure is where you'll have the really bad performance. If you know the exact full pathname of a file, you'll have great performance. An ls or find will be painfully slow on version 3 and 4.</description>
    <pubDate>Tue, 24 Apr 2007 19:17:32 GMT</pubDate>
    <dc:creator>Bill Hassell</dc:creator>
    <dc:date>2007-04-24T19:17:32Z</dc:date>
    <item>
      <title>millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985540#M604671</link>
      <description>Is there any filesystem that supports millions of files per directory?  These are very small files.  Assuming that the filesystem must continue to support the large numbers, how can I alleviate the obvious performance issues?  Does HP, Veritas, Pillar, NetApp, EMC, or anyone have a solution?</description>
      <pubDate>Fri, 20 Apr 2007 06:04:43 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985540#M604671</guid>
      <dc:creator>Joe Odman</dc:creator>
      <dc:date>2007-04-20T06:04:43Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985541#M604672</link>
      <description>The problem with having millions of files in a directory is when you run commands against the directory ie ls or find etc the command will run for an exessive period of time and if you use a wild card the the search ie ls -l * this may fail due to memory.&lt;BR /&gt;&lt;BR /&gt;Other things like backups or filesystem sychronisation may be prolonged due to the number of files that have to be opened and written etc.&lt;BR /&gt;&lt;BR /&gt;As far as I am aware the inode limit is the setting that determines how many files can be in a filesystem as each file is added to the inode table.&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Fri, 20 Apr 2007 06:11:32 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985541#M604672</guid>
      <dc:creator>lawrenzo</dc:creator>
      <dc:date>2007-04-20T06:11:32Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985542#M604673</link>
      <description>what application will be writing these files or accessing them? is there a tuning doc for the application?</description>
      <pubDate>Fri, 20 Apr 2007 06:12:46 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985542#M604673</guid>
      <dc:creator>lawrenzo</dc:creator>
      <dc:date>2007-04-20T06:12:46Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985543#M604674</link>
      <description>Assume that the application cannot change.  The filesystem itself must change.</description>
      <pubDate>Fri, 20 Apr 2007 06:16:40 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985543#M604674</guid>
      <dc:creator>Joe Odman</dc:creator>
      <dc:date>2007-04-20T06:16:40Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985544#M604675</link>
      <description>I have seen the side-effects Lawrenzo is talking about first-hand, and I can tell you it is far from easy to work on a problem within such directories.&lt;BR /&gt;&lt;BR /&gt;Best thing to do (if possible) is to create some sort of hashing-algorithm to put these millions of files in a tree of subdirectories.&lt;BR /&gt;&lt;BR /&gt;For example, given a bunch of files ranging from 'a000000' to 'c999999' you could start by having subdirectories 'a', 'b', and 'c' (each good for one millon files). Within these directories you could then have subdirectories '000' to '999', each holding one thousand files. &lt;BR /&gt;&lt;BR /&gt;A simplified example, of course, but I'd try implementing something like this in order to avoid the complications described above.&lt;BR /&gt;&lt;BR /&gt;Cheers,&lt;BR /&gt;Wout</description>
      <pubDate>Fri, 20 Apr 2007 06:25:13 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985544#M604675</guid>
      <dc:creator>Wouter Jagers</dc:creator>
      <dc:date>2007-04-20T06:25:13Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985545#M604676</link>
      <description>you will have to set the inode limit or at least check the size and determine if this will be reached.&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;setup a script as mentioned to move the files into sub directories or the environment will become unmanageable.&lt;BR /&gt;&lt;BR /&gt;HTH&lt;BR /&gt;#&lt;BR /&gt;&lt;BR /&gt;Chris</description>
      <pubDate>Fri, 20 Apr 2007 06:34:37 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985545#M604676</guid>
      <dc:creator>lawrenzo</dc:creator>
      <dc:date>2007-04-20T06:34:37Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985546#M604677</link>
      <description>Hi Dave:&lt;BR /&gt;&lt;BR /&gt;Divide and conquer.  That said, by using current VxFS (JFS) releases (e.g. 4.1 or later) with the latest version layout and mount options that meet your needs, but offer the best performance, you can probably achieve some gains.&lt;BR /&gt;&lt;BR /&gt;Have a look at thie white paper on JFS performance and tuning:&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://docs.hp.com/en/5576/JFS_Tuning.pdf" target="_blank"&gt;http://docs.hp.com/en/5576/JFS_Tuning.pdf&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;Another good source of mount options as they relate to performance for VxFS filesystems is the manpages for 'mount_vxfs'.  You might find, for instance that mounting with 'noatime' helps speed up your filesystem searches if this is their predominate activity.&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://docs.hp.com/en/B2355-60105/mount_vxfs.1M.html" target="_blank"&gt;http://docs.hp.com/en/B2355-60105/mount_vxfs.1M.html&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;Regards!&lt;BR /&gt;&lt;BR /&gt;...JRF...</description>
      <pubDate>Fri, 20 Apr 2007 06:50:24 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985546#M604677</guid>
      <dc:creator>James R. Ferguson</dc:creator>
      <dc:date>2007-04-20T06:50:24Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985547#M604678</link>
      <description>This is the most appropriate response so far.  We are on JFS layout 4 now, but will look at migrating ot layout 5.  We may also test the noatime. The Netapps WAFL filesystem claims to be better yet, but this is not quantified.  I also found that the ReiserFS Version 4 (SUSE) handles a million files in a directory efficiently.  Still looking for a better HP-UX solution.</description>
      <pubDate>Fri, 20 Apr 2007 08:14:38 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985547#M604678</guid>
      <dc:creator>Joe Odman</dc:creator>
      <dc:date>2007-04-20T08:14:38Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985548#M604679</link>
      <description>I would get a baseball bat and use it on whatever developer or vendor came up with this scheme but if you are looking for solutions that insist that the application be maintained as is, I would say that the only viable alternative is a solid-state disk. Directory searches are linear and it will still require on average n/2 accesses to find a given file; at least with a solid-state disk (backed up automatically and transparently with conventional disks), these searches will be as fast as possible.&lt;BR /&gt;&lt;BR /&gt;You are essentially using the directory as a database - something that it was never intended to do.&lt;BR /&gt;</description>
      <pubDate>Fri, 20 Apr 2007 09:37:54 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985548#M604679</guid>
      <dc:creator>A. Clay Stephenson</dc:creator>
      <dc:date>2007-04-20T09:37:54Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985549#M604680</link>
      <description>I had investigated "content" storage appliances, such as EMC's Celerra several years ago.   They used a diffent approach, called "node based" storage, where the solution is actually a group of small servers, each with their own local storage, a set of master nodes, used to arrange the storage.    EMC claimed that this system could handle millions of files.&lt;BR /&gt;&lt;BR /&gt;Could I ask why?  Could you work out a front-end access script/program that would sort and store programs in separate directies instead?  &lt;BR /&gt;&lt;BR /&gt;-tjh</description>
      <pubDate>Sun, 22 Apr 2007 08:36:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985549#M604680</guid>
      <dc:creator>Thomas J. Harrold</dc:creator>
      <dc:date>2007-04-22T08:36:44Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985550#M604681</link>
      <description>&amp;gt; I had investigated "content" storage appliances, such as EMC's Celerra several years ago. They used a diffent approach, called "node based" storage, where the solution is actually a group of small servers, each with their own local storage, a set of master nodes, used to arrange the storage. EMC claimed that this system could handle millions of files.&lt;BR /&gt;&lt;BR /&gt;The product is called Centera and it is the first successful commercial product using CAS (content addressable storage). Millions of files are trivial -- we have dozens of terabytes of small files on Centeras. NOTE: there is no directory structure at all so you need a database to keep track of the special name for each file. Also, performance is limited and really designed for low volume access such as data archiving -- somewhere between disk arrays and tape silos.&lt;BR /&gt; &lt;BR /&gt;As Clay mentioned, you have a developer problem that sounds suspiciously like a way to avoid buying a real database program (ie, every part number in the company has a small file with data in it). VxFS does not have any practical limitation on inodes since they are&lt;BR /&gt;built dynamically. AS long as you have space, you can add more files.&lt;BR /&gt; &lt;BR /&gt;But don't do searches (ie, ls * or find, etc) and expect anything faster than minute responses. Once the developer tries to make this work, you'll probably get Version 2 of the software where a small database tracks all the files with binary searches, or maybe an hash algorithm...hummm, who knows, you may end up with Version 3 which is a real database.&lt;BR /&gt; &lt;BR /&gt;Note: if you want the best possible performance with millions of files, you MUST install 11.31 as there are specific enhancements for massive directories -- and no, it still won't perform like a real database.</description>
      <pubDate>Sun, 22 Apr 2007 20:56:12 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985550#M604681</guid>
      <dc:creator>Bill Hassell</dc:creator>
      <dc:date>2007-04-22T20:56:12Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985551#M604682</link>
      <description>well, we can all agree that the application is a prime example of bad design.&lt;BR /&gt;&lt;BR /&gt;perhaps, reiserfs can handle these things (but that is not available on hpux), it is not only the filesystem itself that has to be effecient with it.&lt;BR /&gt;&lt;BR /&gt;try as hard as you can to get this stuff into a database, that is what these things are made for and what they do best. they will outperform any filesystem for sure.</description>
      <pubDate>Mon, 23 Apr 2007 01:40:42 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985551#M604682</guid>
      <dc:creator>dirk dierickx</dc:creator>
      <dc:date>2007-04-23T01:40:42Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985552#M604683</link>
      <description>Thank you all for your responses.  I will look into Celerra.  &lt;BR /&gt;&lt;BR /&gt;Bill, I assume you are referring to layout 5 of vxfs 3.5?  We are installing it for testing now.&lt;BR /&gt;&lt;BR /&gt;I have also found XFS, developed by SGI, and ported to Suse.  XFS can also support a large number of files in a directory.  We will be testing this also.&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Mon, 23 Apr 2007 04:49:30 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985552#M604683</guid>
      <dc:creator>Joe Odman</dc:creator>
      <dc:date>2007-04-23T04:49:30Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985553#M604684</link>
      <description>Bill is correct, it's "centerra", not "celerra".  Celerra is an NFS head for an EMC storage array.  I've worked with both, and always got the names confused.  :)&lt;BR /&gt;&lt;BR /&gt;-tjh</description>
      <pubDate>Mon, 23 Apr 2007 08:36:48 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985553#M604684</guid>
      <dc:creator>Thomas J. Harrold</dc:creator>
      <dc:date>2007-04-23T08:36:48Z</dc:date>
    </item>
    <item>
      <title>Re: millions of files per directory</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985554#M604685</link>
      <description>That's correct. The version 5 directory structure (and associated subsystem code) is much more efficient with directory scans. Note that VxFS (any version) can have essentially unlimited files. What you do to manipulate the the structure is where you'll have the really bad performance. If you know the exact full pathname of a file, you'll have great performance. An ls or find will be painfully slow on version 3 and 4.</description>
      <pubDate>Tue, 24 Apr 2007 19:17:32 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/millions-of-files-per-directory/m-p/3985554#M604685</guid>
      <dc:creator>Bill Hassell</dc:creator>
      <dc:date>2007-04-24T19:17:32Z</dc:date>
    </item>
  </channel>
</rss>

