[Beowulf] automount on high ports
Gerry Creager
gerry.creager at tamu.edu
Wed Jul 2 07:09:34 PDT 2008
Scott Atchley wrote:
> On Jul 2, 2008, at 7:22 AM, Carsten Aulbert wrote:
>
>> Bogdan Costescu wrote:
>>>
>>> Have you considered using a parallel file system ?
>>
>> We looked a bit into a few, but would love to get any input from anyone
>> on that. What we found so far was not really convincing, e.g. glusterFS
>> at that time was not really stable, lustre was too easy to crash - at l
>> east at that time, ...
>
> Hi Carsten,
>
> I have not looked at GlusterFS at all. I have worked with Lustre and
> PVFS2 (I wrote the shims to allow them to run on MX).
>
> Although I believe Lustre's robustness is very good these days, I do not
> believe that it will not work in your setting. I think that they
> currently do not recommend mounting a client on a node that is also
> working as a server as you are doing with NFS. I believe it is due to
> memory contention leading to deadlock.
Lustre is good enough that it's the parallel FS at TACC for the Ranger
cluster. And, I've had no real problems as a user thereof. We're
brining up glustre on our new cluster here ( <flamebait> CentOS/RHEL5,
not debian </flamebait>). We looked at zfs but didn't have sufficient
experience to go that path.
> PVFS2 does, however, support your scenario where each node is a server
> and can be mounted locally as well. PVFS2 servers run in userspace and
> can be easily debugged. If you are using MPI-IO, it integrates nicely as
> well. Even so, keep in mind that using each node as a server will
> consume network resources and will compete with MPI communications.
Someone at NCAR recently suggested we review PVFS2. I'm gonna do it as
soon as I get a free moment on vacation.
--
Gerry Creager -- gerry.creager at tamu.edu
Texas Mesonet -- AATLT, Texas A&M University
Cell: 979.229.5301 Office: 979.458.4020 FAX: 979.862.3983
Office: 1700 Research Parkway Ste 160, TAMU, College Station, TX 77843
More information about the Beowulf
mailing list