I’m new to GPU’s, investigating them for some scientific computation, and they look very exciting! I’ve been given a sudden deadline to get an estimate of whether GPU processing will be sufficient for a project proposal that’s being prepared. Is there somewhere I can read about maximum theoretical flops for the top GPU chip(s)? Or somewhere to read about how to estimate this? Thanks for any advice, I plan to come back and look at all this more thoroughly when I have the time.
I saw in the FAQ that acceleration of data parallel projects can range from 10x-200x. That’s encouraging, but I need to find out how close to 100x we may get. At least with a best estimate, I imagine implementation will affect a lot).
The algorithm needing acceleration is highly data parallel, and with a minimally divergent kernal (I think that’s the term, it has few logic branches, just cruching a massive step-wise integration).