[Beowulf] glusterfs and openmpi/mpich problems
Gerry Creager
gerry.creager at tamu.edu
Thu Jan 8 10:25:55 PST 2009
We've been working with gluster of late, on our high throughput cluster
(126 nodes, gigabit connected). We did some tweaking recently, and now,
my test code, an instance of WRF on 128 cores, just sorta dies.
More specifically, it takes 19 minutes to write the first 403MB file to
disk, while various tasks are mindlessly using CPU time, but only the
initial output file appears to get written.
Does anyone have any history with gluster who might be willing to offer
some help/hints?
Thanks, Gerry
More information about the Beowulf
mailing list