[Beowulf] new mpich/pvfs2
maillists at gauckler.ch
Sun Jun 12 07:51:23 PDT 2005
Dear list, dear mpich/pvfs2 developers,
thank you very much for the Release of mpich2 version 1.0.2 together
with PVFS2 1.1.0. We noticed that the performance increased and is now
independent of the cb_buffer_size. We are experiencing datarates of up
to 220 MB/s (which was previously only achievable for certain sizes of
cb_buffer_size) measured by coll_perf. Also memory utilisation when
flattening the MPI_Type_create_subarray came down (from in our case
To write a 1.6 GB file from 2 processes in a subarray-sliced manner to
disk and read it again, 191 sec. are needed (~17 MB/s, see ). Doing
the same test with 4 processors (3.2 GB of data) unfortunately is not
possible since the program crashes. Before the upgrade of mpich/pvfs2
the programm run with 4 processors, so the problem is unlikely in the
program (see ) itself. The debugger stops the program in a PVFS
call (see )
We are happy for any hint how to avoid this or any
timings/confirmation from people who are willing to test our little
program on their system.
Thank you for your time!
We are running Debian on Intel Xeon:
Linux 2.6.8-2-686-smp #1 SMP Thu May 19 17:27:55 JST 2005 i686
time mpdrun -1 -np 2 ./mpicube2
 --> Attachement "mpicube2_mpich_v102_pvfs2_v110.cpp"
(See attached file: mpicube2_mpich_v102_pvfs2_v110.cpp)
 --> http://www.parallelsystems.ch/mpicube2_debug.png
-------------- next part --------------
A non-text attachment was scrubbed...
Size: 9890 bytes
Desc: not available
Url : http://www.scyld.com/pipermail/beowulf/attachments/20050612/444ee61e/mpicube2_mpich_v102_pvfs2_v110.bin
More information about the Beowulf