<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: CPU STALLS in Operating System - HP-UX</title>
    <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814298#M268865</link>
    <description>Thanks clay,&lt;BR /&gt;&lt;BR /&gt;in fact, cache misses are only 12%&lt;BR /&gt;&lt;BR /&gt;Here is a report.........What do you suggest?&lt;BR /&gt;&lt;BR /&gt;Please guide.&lt;BR /&gt;&lt;BR /&gt;Anurag&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;L1 data cache miss percentage:&lt;BR /&gt;&lt;BR /&gt;Sampling Specification&lt;BR /&gt;  Sampling event:             DATA_EAR_EVENTS&lt;BR /&gt;  Sampling period:            10000 events&lt;BR /&gt;  Sampling period variation:  500 (5.00% of sampling period)&lt;BR /&gt;  Sampling counter privilege: user (user-space sampling)&lt;BR /&gt;  Data granularity:           16 bytes&lt;BR /&gt;  Number of samples:          452&lt;BR /&gt;  Data sampled:               Data cache miss&lt;BR /&gt;&lt;BR /&gt;Data Cache Metrics Summed for Entire Run&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;                         PLM&lt;BR /&gt;Event Name              U..K  TH          Count&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;DATA_REFERENCES         x___   0      447196040&lt;BR /&gt;L1D_READS               x___   0      312349645&lt;BR /&gt;L1D_READ_MISSES.ALL     x___   0       37692146&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;PLM: Privilege Level Mask&lt;BR /&gt;     U/K = user/kernel levels (U: level 3, K: level 0)&lt;BR /&gt;     The intermediate levels (1, 2) are unused on HP-UX or Linux&lt;BR /&gt;     x : the metric is measured at the given level (_ : not measured)&lt;BR /&gt;TH: event THreshold, determines the event counter behavior,&lt;BR /&gt;    TH == 0 : counter += event_count_in_cycle&lt;BR /&gt;    TH  &amp;gt; 0 : counter += (event_count_in_cycle &amp;gt;= threshold ? 1 : 0)&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;L1 data cache miss percentage:&lt;BR /&gt;  12.07 = 100 * (L1D_READ_MISSES.ALL / L1D_READS)&lt;BR /&gt;&lt;BR /&gt;Percent of data references accessing L1 data cache:&lt;BR /&gt;  69.85 = 100 * (L1D_READS / DATA_REFERENCES)&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;&lt;BR /&gt;Load Module Summary&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;% Total                                        Avg.&lt;BR /&gt; Dcache  Cumulat        Sampled       Dcache  Dcache&lt;BR /&gt;Latency    % of          Dcache      Latency  Laten.&lt;BR /&gt; Cycles    Total         Misses       Cycles  Cycles   Load Module&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt; 95.97    95.97             365        13715    37.6   dld.so&lt;BR /&gt;  1.66    97.63              31          237     7.6   libunwind.so.1&lt;BR /&gt;  0.85    98.48              17          122     7.2   libpthread.so.1&lt;BR /&gt;  0.82    99.30              10          117    11.7   libncursesw.so&lt;BR /&gt;  0.38    99.69               6           55     9.2   librtc.sl&lt;BR /&gt;  0.13    99.82               3           19     6.3   liborb_r.so&lt;BR /&gt;  0.13    99.95               2           19     9.5   libc.so.1&lt;BR /&gt;  0.05   100.00               1            7     7.0   libCsup.so.1&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;100.00   100.00             435        14291    32.9   Total&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;&lt;BR /&gt;Function Summary&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;% Total                                        Avg.&lt;BR /&gt; Dcache  Cumulat        Sampled       Dcache  Dcache&lt;BR /&gt;Latency    % of          Dcache      Latency  Laten.&lt;BR /&gt; Cycles    Total         Misses       Cycles  Cycles  Function                          File&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;  0.76     0.76               9          108    12.0  libncursesw.so::__milli_memcmp&lt;BR /&gt;  0.40     1.15               7           57     8.1  libunwind.so.1::uwx_step          uwx_step.c&lt;BR /&gt;  0.29     1.44               6           41     6.8  libpthread.so.1::*unnamed@0x4042(920-cc0)*  mutex.c&lt;BR /&gt;  0.25     1.69               4           36     9.0  libunwind.so.1::uwx_get_frame_info  uwx_step.c&lt;BR /&gt;  0.19     1.88               3           27     9.0  libpthread.so.1::pthread_setcancelstate  cancel.c&lt;BR /&gt;  0.19     2.07               3           27     9.0  libunwind.so.1::uwx_reclaim_scoreboards  uwx_scoreboard.c&lt;BR /&gt;  0.17     2.24               4           24     6.0  libunwind.so.1::uwx_decode_prologue  uwx_uinfo.c&lt;BR /&gt;  0.15     2.39               1           21    21.0  { STUB }-&amp;gt;libunwind.so.1::uwx_reset_str_pool&lt;BR /&gt;  0.14     2.53               4           20     5.0  libpthread.so.1::pthread_mutex_lock  mutex.c&lt;BR /&gt;  0.13     2.65               2           18     9.0  libpthread.so.1::pthread_mutex_unlock  mutex.c&lt;BR /&gt;  0.13     2.78               2           18     9.0  librtc.sl::rtc_split_special_region  infrtc.c&lt;BR /&gt;  0.11     2.89               2           16     8.0  libpthread.so.1::ENTER_PTHREAD_LIBRARY_FUNC  pthread.c&lt;BR /&gt;  0.10     2.99               3           15     5.0  libunwind.so.1::uwx_search_utable32  uwx_utable.c&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;[Minimum function entries: 5, percent cutoff: 0.10, cumulative percent cutoff: 100.00]&lt;BR /&gt;&lt;BR /&gt;Function Details&lt;BR /&gt;----------------------------------------------------------------------&lt;BR /&gt;% Total                               Avg.&lt;BR /&gt; Dcache        Sampled       Dcache  Dcache         Line|&lt;BR /&gt;Latency         Dcache      Latency  Laten.         Slot|  &amp;gt;Statement|&lt;BR /&gt; Cycles         Misses       Cycles  Cycles    Col,Offset  Instruction&lt;BR /&gt;----------------------------------------------------------------------&lt;BR /&gt;[Cutoffs excluded all entries (minimum: 0; percent: 1.00; cumulative percent: 100.00;)]&lt;BR /&gt;</description>
    <pubDate>Wed, 28 Jun 2006 12:32:54 GMT</pubDate>
    <dc:creator>Anurag_7</dc:creator>
    <dc:date>2006-06-28T12:32:54Z</dc:date>
    <item>
      <title>CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814291#M268858</link>
      <description>We have our 32 bit application running on Itanium.&lt;BR /&gt;&lt;BR /&gt;The application response is a cause of concern.&lt;BR /&gt;&lt;BR /&gt;% of cycles lost due to CPU stalls is 71.45.&lt;BR /&gt;&lt;BR /&gt;Analysis of a process using caliper suggests that the CPU stall is very high. Below is the report.&lt;BR /&gt;&lt;BR /&gt;Is there a way to overcome this?&lt;BR /&gt;&lt;BR /&gt;Please guide.&lt;BR /&gt;&lt;BR /&gt;Anurag&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;HP Caliper Total CPU Event Counts Report for PkMS&lt;BR /&gt;================================================================================&lt;BR /&gt;&lt;BR /&gt;Target Application&lt;BR /&gt;  Program:                    /u03/wmpso/CMLACARE/opt/app/wmprd/bin/Lttr&lt;BR /&gt;  Invocation:                 /u03/wmpso/CMLACARE/opt/app/wmprd/bin/Lttr&lt;BR /&gt;  Process ID:                 3456 (started by Caliper)&lt;BR /&gt;  Start time:                 09:43:11 PM&lt;BR /&gt;  End time:                   09:43:16 PM&lt;BR /&gt;  Termination Status:         0&lt;BR /&gt;  Last modified:              May 05, 2006 at 09:26 PM&lt;BR /&gt;  Memory model:               ILP32&lt;BR /&gt;  Main module text page size: default&lt;BR /&gt;&lt;BR /&gt;Processor Information&lt;BR /&gt;  Machine name:         itanium1&lt;BR /&gt;  Number of processors: 4&lt;BR /&gt;  Processor type:       Itanium2 6M&lt;BR /&gt;  Processor speed:      1300 MHz&lt;BR /&gt;&lt;BR /&gt;Target Execution Time&lt;BR /&gt;  Real time:   4.915 seconds&lt;BR /&gt;  User time:   1.757 seconds&lt;BR /&gt;  System time: 0.228 seconds&lt;BR /&gt;&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;                         PLM&lt;BR /&gt;Event Name              U..K  TH          Count&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;BACK_END_BUBBLE.ALL     x___   0     1551927751&lt;BR /&gt;CPU_CYCLES              x___   0     2171937631&lt;BR /&gt;IA64_INST_RETIRED       x___   0     2094967607&lt;BR /&gt;NOPS_RETIRED            x___   0      403213401&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;PLM: Privilege Level Mask&lt;BR /&gt;     U/K = user/kernel levels (U: level 3, K: level 0)&lt;BR /&gt;     The intermediate levels (1, 2) are unused on HP-UX or Linux&lt;BR /&gt;     x : the metric is measured at the given level (_ : not measured)&lt;BR /&gt;TH: event THreshold, determines the event counter behavior,&lt;BR /&gt;    TH == 0 : counter += event_count_in_cycle&lt;BR /&gt;    TH  &amp;gt; 0 : counter += (event_count_in_cycle &amp;gt;= threshold ? 1 : 0)&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;% of Cycles lost due to stalls (lower is better):&lt;BR /&gt;  100 * BACK_END_BUBBLE.ALL / CPU_CYCLES = 71.45&lt;BR /&gt;&lt;BR /&gt;Effective CPI (lower is better):&lt;BR /&gt;  CPU_CYCLES / (IA64_INST_RETIRED - NOPS_RETIRED) = 1.2838&lt;BR /&gt;&lt;BR /&gt;Effective CPI during unstalled execution (lower is better):&lt;BR /&gt;  (CPU_CYCLES - BACK_END_BUBBLE.ALL) / (IA64_INST_RETIRED - NOPS_RETIRED) = 0.3665&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;</description>
      <pubDate>Wed, 28 Jun 2006 11:37:23 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814291#M268858</guid>
      <dc:creator>Anurag_7</dc:creator>
      <dc:date>2006-06-28T11:37:23Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814292#M268859</link>
      <description>Shalom,&lt;BR /&gt;&lt;BR /&gt;This is a bit strange, but I think I'd have the hardware checked.&lt;BR /&gt;&lt;BR /&gt;First use cstm mstm or xstm and run the cpu excercizes. Any failures, HP needs to replace.&lt;BR /&gt;&lt;BR /&gt;Then call in the HP Hardware team.&lt;BR /&gt;&lt;BR /&gt;SEP</description>
      <pubDate>Wed, 28 Jun 2006 11:58:03 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814292#M268859</guid>
      <dc:creator>Steven E. Protter</dc:creator>
      <dc:date>2006-06-28T11:58:03Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814293#M268860</link>
      <description>Hi Shalom,&lt;BR /&gt;&lt;BR /&gt;Thanks a lot for your reply.&lt;BR /&gt;&lt;BR /&gt;Can you please let me know as to how to use these commands and what kind of output should i collect and analyze?&lt;BR /&gt;&lt;BR /&gt;Thanks a lot&lt;BR /&gt;&lt;BR /&gt;Anurag</description>
      <pubDate>Wed, 28 Jun 2006 12:02:59 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814293#M268860</guid>
      <dc:creator>Anurag_7</dc:creator>
      <dc:date>2006-06-28T12:02:59Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814294#M268861</link>
      <description>It looks like there is a big chunk of branch misprediction occurring which is starving the processor for instructions. Since you mention 32-bit application, I assume that means your are running a PA-RISC application under the ARIES emulator. Do you have the option of porting the application to native code?</description>
      <pubDate>Wed, 28 Jun 2006 12:07:31 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814294#M268861</guid>
      <dc:creator>A. Clay Stephenson</dc:creator>
      <dc:date>2006-06-28T12:07:31Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814295#M268862</link>
      <description>Its not a PA-RISC binary :-)&lt;BR /&gt;&lt;BR /&gt;Its a  ELF-32 executable object file - IA64&lt;BR /&gt;</description>
      <pubDate>Wed, 28 Jun 2006 12:15:39 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814295#M268862</guid>
      <dc:creator>Anurag_7</dc:creator>
      <dc:date>2006-06-28T12:15:39Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814296#M268863</link>
      <description>71.45 is a high number.  Greater then 50 is unacceptable.  But this all is an indication of a poorly written program.  (loops in loops vs. straight line code.) Check for cache misses with caliper.&lt;BR /&gt;&lt;BR /&gt;caliper icache -o reports/icachem.txt ./matmul&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://docs.hp.com/en/5991-5499/ch02s04.html" target="_blank"&gt;http://docs.hp.com/en/5991-5499/ch02s04.html&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;"...A hot spot is an instruction or set of instructions that has a higher execution count than most other instructions in a program. For example, code that is inside a loop inside a loop inside a loop will likely be executed more times than straight-line code. Usually the â  hotnessâ   is measured with CPU cycles, but it could also be measured with metrics such as cache misses...."&lt;BR /&gt;&lt;BR /&gt;this doc. makes some patching suggestings as well as increasing page size:&lt;BR /&gt;&lt;BR /&gt;&lt;A href="http://h21007.www2.hp.com/dspp/files/unprotected/hpux/Top_Ten_Perf_Tips.pdf" target="_blank"&gt;http://h21007.www2.hp.com/dspp/files/unprotected/hpux/Top_Ten_Perf_Tips.pdf&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;There's also a lot of reference to the latest pthread patch.  See page 12 MxN v. 1x1.</description>
      <pubDate>Wed, 28 Jun 2006 12:21:44 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814296#M268863</guid>
      <dc:creator>Michael Steele_2</dc:creator>
      <dc:date>2006-06-28T12:21:44Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814297#M268864</link>
      <description>In that case, the most likely culprit is poorly written code so that there are tons of cache misses. I would almost certainly rule out any sort of hardware problem since I assume this hardware doesn't know how to just pick on your application. If you see widespread performance degradation then I would reconsider but if the poor performance is limited to your application then ...</description>
      <pubDate>Wed, 28 Jun 2006 12:27:57 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814297#M268864</guid>
      <dc:creator>A. Clay Stephenson</dc:creator>
      <dc:date>2006-06-28T12:27:57Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814298#M268865</link>
      <description>Thanks clay,&lt;BR /&gt;&lt;BR /&gt;in fact, cache misses are only 12%&lt;BR /&gt;&lt;BR /&gt;Here is a report.........What do you suggest?&lt;BR /&gt;&lt;BR /&gt;Please guide.&lt;BR /&gt;&lt;BR /&gt;Anurag&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;L1 data cache miss percentage:&lt;BR /&gt;&lt;BR /&gt;Sampling Specification&lt;BR /&gt;  Sampling event:             DATA_EAR_EVENTS&lt;BR /&gt;  Sampling period:            10000 events&lt;BR /&gt;  Sampling period variation:  500 (5.00% of sampling period)&lt;BR /&gt;  Sampling counter privilege: user (user-space sampling)&lt;BR /&gt;  Data granularity:           16 bytes&lt;BR /&gt;  Number of samples:          452&lt;BR /&gt;  Data sampled:               Data cache miss&lt;BR /&gt;&lt;BR /&gt;Data Cache Metrics Summed for Entire Run&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;                         PLM&lt;BR /&gt;Event Name              U..K  TH          Count&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;DATA_REFERENCES         x___   0      447196040&lt;BR /&gt;L1D_READS               x___   0      312349645&lt;BR /&gt;L1D_READ_MISSES.ALL     x___   0       37692146&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;PLM: Privilege Level Mask&lt;BR /&gt;     U/K = user/kernel levels (U: level 3, K: level 0)&lt;BR /&gt;     The intermediate levels (1, 2) are unused on HP-UX or Linux&lt;BR /&gt;     x : the metric is measured at the given level (_ : not measured)&lt;BR /&gt;TH: event THreshold, determines the event counter behavior,&lt;BR /&gt;    TH == 0 : counter += event_count_in_cycle&lt;BR /&gt;    TH  &amp;gt; 0 : counter += (event_count_in_cycle &amp;gt;= threshold ? 1 : 0)&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;L1 data cache miss percentage:&lt;BR /&gt;  12.07 = 100 * (L1D_READ_MISSES.ALL / L1D_READS)&lt;BR /&gt;&lt;BR /&gt;Percent of data references accessing L1 data cache:&lt;BR /&gt;  69.85 = 100 * (L1D_READS / DATA_REFERENCES)&lt;BR /&gt;-----------------------------------------------&lt;BR /&gt;&lt;BR /&gt;Load Module Summary&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;% Total                                        Avg.&lt;BR /&gt; Dcache  Cumulat        Sampled       Dcache  Dcache&lt;BR /&gt;Latency    % of          Dcache      Latency  Laten.&lt;BR /&gt; Cycles    Total         Misses       Cycles  Cycles   Load Module&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt; 95.97    95.97             365        13715    37.6   dld.so&lt;BR /&gt;  1.66    97.63              31          237     7.6   libunwind.so.1&lt;BR /&gt;  0.85    98.48              17          122     7.2   libpthread.so.1&lt;BR /&gt;  0.82    99.30              10          117    11.7   libncursesw.so&lt;BR /&gt;  0.38    99.69               6           55     9.2   librtc.sl&lt;BR /&gt;  0.13    99.82               3           19     6.3   liborb_r.so&lt;BR /&gt;  0.13    99.95               2           19     9.5   libc.so.1&lt;BR /&gt;  0.05   100.00               1            7     7.0   libCsup.so.1&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;100.00   100.00             435        14291    32.9   Total&lt;BR /&gt;------------------------------------------------------------------&lt;BR /&gt;&lt;BR /&gt;Function Summary&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;% Total                                        Avg.&lt;BR /&gt; Dcache  Cumulat        Sampled       Dcache  Dcache&lt;BR /&gt;Latency    % of          Dcache      Latency  Laten.&lt;BR /&gt; Cycles    Total         Misses       Cycles  Cycles  Function                          File&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;  0.76     0.76               9          108    12.0  libncursesw.so::__milli_memcmp&lt;BR /&gt;  0.40     1.15               7           57     8.1  libunwind.so.1::uwx_step          uwx_step.c&lt;BR /&gt;  0.29     1.44               6           41     6.8  libpthread.so.1::*unnamed@0x4042(920-cc0)*  mutex.c&lt;BR /&gt;  0.25     1.69               4           36     9.0  libunwind.so.1::uwx_get_frame_info  uwx_step.c&lt;BR /&gt;  0.19     1.88               3           27     9.0  libpthread.so.1::pthread_setcancelstate  cancel.c&lt;BR /&gt;  0.19     2.07               3           27     9.0  libunwind.so.1::uwx_reclaim_scoreboards  uwx_scoreboard.c&lt;BR /&gt;  0.17     2.24               4           24     6.0  libunwind.so.1::uwx_decode_prologue  uwx_uinfo.c&lt;BR /&gt;  0.15     2.39               1           21    21.0  { STUB }-&amp;gt;libunwind.so.1::uwx_reset_str_pool&lt;BR /&gt;  0.14     2.53               4           20     5.0  libpthread.so.1::pthread_mutex_lock  mutex.c&lt;BR /&gt;  0.13     2.65               2           18     9.0  libpthread.so.1::pthread_mutex_unlock  mutex.c&lt;BR /&gt;  0.13     2.78               2           18     9.0  librtc.sl::rtc_split_special_region  infrtc.c&lt;BR /&gt;  0.11     2.89               2           16     8.0  libpthread.so.1::ENTER_PTHREAD_LIBRARY_FUNC  pthread.c&lt;BR /&gt;  0.10     2.99               3           15     5.0  libunwind.so.1::uwx_search_utable32  uwx_utable.c&lt;BR /&gt;--------------------------------------------------------------------------------------------&lt;BR /&gt;[Minimum function entries: 5, percent cutoff: 0.10, cumulative percent cutoff: 100.00]&lt;BR /&gt;&lt;BR /&gt;Function Details&lt;BR /&gt;----------------------------------------------------------------------&lt;BR /&gt;% Total                               Avg.&lt;BR /&gt; Dcache        Sampled       Dcache  Dcache         Line|&lt;BR /&gt;Latency         Dcache      Latency  Laten.         Slot|  &amp;gt;Statement|&lt;BR /&gt; Cycles         Misses       Cycles  Cycles    Col,Offset  Instruction&lt;BR /&gt;----------------------------------------------------------------------&lt;BR /&gt;[Cutoffs excluded all entries (minimum: 0; percent: 1.00; cumulative percent: 100.00;)]&lt;BR /&gt;</description>
      <pubDate>Wed, 28 Jun 2006 12:32:54 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814298#M268865</guid>
      <dc:creator>Anurag_7</dc:creator>
      <dc:date>2006-06-28T12:32:54Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814299#M268866</link>
      <description>I more concerned about instruction cache misses. You may well benefit from using profile based optimization.&lt;BR /&gt;&lt;BR /&gt;It's done like this:&lt;BR /&gt;&lt;BR /&gt;aCC +Oprofile=collect -O sample.C -o sample.exe                            // Compile to instrumented executable. sample.exe &amp;lt; input.file     // Collect execution profile data. aCC +Oprofile=use -O sample.C -o sample.exe                             // Recompile with optimization&lt;BR /&gt;&lt;BR /&gt;Profiled base optimization will make much better decisions in laying the code because it uses statistics gathered during an actual execution.</description>
      <pubDate>Wed, 28 Jun 2006 14:07:25 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814299#M268866</guid>
      <dc:creator>A. Clay Stephenson</dc:creator>
      <dc:date>2006-06-28T14:07:25Z</dc:date>
    </item>
    <item>
      <title>Re: CPU STALLS</title>
      <link>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814300#M268867</link>
      <description>Oh, and because this is UNIX, and because as a general rule, UNIX code tends to be i/o bound rather than actually CPU bound (intense analysis programs are the exception) make sure that you are not spending tons of time optimizing code when the CPU may only be a small component of the actual bottleneck.</description>
      <pubDate>Wed, 28 Jun 2006 14:10:29 GMT</pubDate>
      <guid>https://community.hpe.com/t5/operating-system-hp-ux/cpu-stalls/m-p/3814300#M268867</guid>
      <dc:creator>A. Clay Stephenson</dc:creator>
      <dc:date>2006-06-28T14:10:29Z</dc:date>
    </item>
  </channel>
</rss>

