[Beowulf] updated GPU-HMMer/mpiHMMer bits

Joe Landman landman at scalableinformatics.com
Sun Feb 8 21:11:42 PST 2009


This time compared to a loaner AMD shanghai 2.3 GHz (same one I did the 
rzf tests on a few weeks ago).

 From the mpihmmer mailing list:

---
A new release of GPU-HMMER is available at www.mpihmmer.org.  The most
notable change in the new code is support for multi-GPU systems.  We
have tested the current GPU-HMMER with up to 3 GPUs, and have achieved
over 100x speedup with sufficiently large HMMs.  A few bug fixes have
been applied as well, so I would encourage users to update.  Users who
update should be aware that several command line options have changed,
and should check the GPU-HMMER user guide for details
(http://www.mpihmmer.org/userguideGPUHMMER.htm).

While the system requirements haven't changed from the last version,
users who intend to use multiple GPUs should be aware that they will
need a substantial amount of system memory in order to do so.  The 3 GPU
system I've been using has 16GB RAM.  This is probably a bit overkill,
but 8GB or so would probably be appropriate.

As always, any comments, bug reports, etc. are welcome.

best regards,
JP Walters

---

Updated mpiHMMer results were shown at SC08, the graph I saw showed 
maxing out about 180x over a single thread, though I think the parallel 
IO version can scale higher still.

Cudos to JP and the team for doing a great job on this!

-- 
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics LLC,
email: landman at scalableinformatics.com
web  : http://www.scalableinformatics.com
        http://jackrabbit.scalableinformatics.com
phone: +1 734 786 8423 x121
fax  : +1 866 888 3112
cell : +1 734 612 4615



More information about the Beowulf mailing list