[Beowulf] high %system utiliziation on infiniband nodes

Walid walid.shaari at gmail.com
Sat Jul 26 08:33:11 PDT 2008


Hi,

I have two nodes Interconnected using Infiniband, and using Intel-MPI over
dapl1.2.7 from OFED 1.3.1 compiled localy on the same build, when there is
interconnect communication i can see on one of the nodes that i monitoring
have a high cpu utiliztion (%system) that exceeds 60%. the mpi job is
helloworld/pallas runing over two nodes, 8 cores each (16 processes in
total)

a snapshot of mpstat -P ALL on one node
06:22:20 PM  CPU   %user   %nice %system %iowait    %irq   %soft   %idle
intr/s
06:22:22 PM  all   30.25    0.00   69.75    0.00    0.00    0.00    0.00
1768.50
06:22:22 PM    0   30.00    0.00   70.00    0.00    0.00    0.00    0.00
566.50
06:22:22 PM    1   30.50    0.00   69.00    0.00    0.00    0.00    0.00
201.00
06:22:22 PM    2   30.50    0.00   69.50    0.00    0.00    0.00
0.00      0.00
06:22:22 PM    3   29.50    0.00   70.50    0.00    0.00    0.00
0.00      0.00
06:22:22 PM    4   28.50    0.00   71.00    0.00    0.00    0.00
0.00      0.00
06:22:22 PM    5   30.00    0.00   70.00    0.00    0.00    0.00
0.00      0.00
06:22:22 PM    6   31.00    0.00   69.50    0.00    0.00    0.00    0.00
1000.50
06:22:22 PM    7   32.00    0.00   68.00    0.00    0.00    0.00
0.00      0.00

now i get the same behaviour on RHEL5.0/5.1 and RHEL4.6, using Infiniband or
ethernet, so is this normal, to me it does not, or at least i have never
seen such behaviour before? the node is a DELL PE1950

regards

Walid
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://www.scyld.com/pipermail/beowulf/attachments/20080726/998be018/attachment.html


More information about the Beowulf mailing list