[Beowulf] Nvidia, cuda, tesla and... where's my double floating point?
Vincent Diepeveen
diep at xs4all.nl
Sun Jun 15 04:31:12 PDT 2008
Seems the next CELL is 100% confirmed double precision.
Yet if you look back in history, Nvidia promises on this can be found
years back.
The only problem with hardware like Tesla is that it is rather hard to
get technical information; like which instructions does Tesla support
in hardware?
This is crucial to know in order to speedup your code.
It is already tough to get realworld codes on GPU's faster than at
cpu's.
The equivalent CPU code has been optimized real bigtime,
knowing everything about hardware.
How fast is latency from RAM when all 128 SP's are busy with that?
Nvidia gives out zero information and doesn't support anyone either
for this.
That has to change in order to get GPU calculations more into
mainstream.
When i calculate on paper for some applications, a GPU can be
potentially
factor 4-8 faster than a standard quadcore 2.4ghz is right now.
Getting that performance out of the GPU is more than a fulltime task
however,
without having indepth technical hardware data on the GPU.
Vincent
On May 5, 2008, at 9:40 PM, John Hearns wrote:
> On Fri, 2008-05-02 at 14:05 +0100, Ricardo Reis wrote:
>> Does anyone knows if/when there will be double floating point on
>> those
>> little toys from nvidia?
>>
>>
> Ricardo,
> I think CUDA is a gret concept, and am starting to work with it at
> home.
> I recently went to a talk by David Kirk, as part of the "world tour".
> I think the answer to your question is Real Soon Now.
>
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org
> To change your subscription (digest mode or unsubscribe) visit
> http://www.beowulf.org/mailman/listinfo/beowulf
>
More information about the Beowulf
mailing list