[Beowulf] Lowered latency with multi-rail IB?
Håkon Bugge
hbugge at platform.com
Fri Mar 27 11:27:15 PDT 2009
, Håkon
On Mar 27, 2009, at 19:09 , Joshua mora acosta wrote:
> So a way to quantify if multirail helps on network latency driven
> workloads
> there should be a sinthetic benchmark that can be built to show off
> the impact
> of balancing these requests among multiple HCAs bound to different
> network
> paths or core pairs like an all-to-[all,gather,scatter] or barrier
> benchmark
> and in theory observe half the total latency of that overall
> communication.
> So I think multirail will reduce latency of sinchronizations
> (collective
> calls) but not to latency driven point to point communications.
>
Well, you're wrong. As stated, we do see speedup due to increased
message rate on apps not using collectives. As to a benchmark, and all-
to-all with _many_ processes per node will show you this. ($MPI_HOME/
examples/bin/mpi_msg_rate in our distro).
I would actually claim the opposite; it will _not_ help on most
collective operations, because they perform SMP optimizations and one
process sends and receives on behalf of the other processes.
Uncorrelated messages from many processes on a single node will, on
the other hand, take advantages of the accumulated increased message
rate provided by multiple HCAs.
Håkon
> Joshua
>
> ------ Original Message ------
> Received: 12:37 PM CDT, 03/27/2009
> From: Håkon Bugge <hbugge at platform.com>
> To: Craig Tierney <Craig.Tierney at noaa.gov>Cc: Joshua mora acosta
> <joshua_mora at usa.net>, DPHURST at uncg.edu, beowulf at beowulf.org
> Subject: Re: [Beowulf] Lowered latency with multi-rail IB?
>
>> On Mar 27, 2009, at 18:20 , Craig Tierney wrote:
>>
>>> What about using multi-rail to increase message rate? That isn't
>>> the same as latency, but if you put messages on both wires you
>>> should get more.
>>
>> Exactly why we saw almost 2x speedup on message rate (latency)
>> sensitive apps using Platform MPI. We call the technique alternating;
>> any communicating peer will use a HCA or port, but a single MPI
>> process will alternate between different HCAs (or ports), depending
>> on
>> which peer he communicates with.
>>
>>
>> Håkon
>
>
>
More information about the Beowulf
mailing list