An email from my friendly NVIDIA rep called my attention to this recent blog post from Andy Keane, head GPU honcho at NVIDIA. In the post, Andy pounds Intel soundly for presenting a paper titled “Debunking the 100x GPU vs. CPU Myth” which, in its abstract, asserts that an older NVIDIA GPU (the GTX280) is only 2.5x faster than Intel’s most current quad-core i7-960.

Intel does a very scholarly job in the paper of laying out their benchmarks, methodology, and results. But it makes one wonder if they – perhaps – could have, well… cherry-picked the benchmarks in order to put the best face on it? I’m sure it’s hard for any of us to imagine this being the case, but the question needs to at least be asked. Right?

For my part, I’ve talked to quite a few real-world folks who have seen 5x, 10x, 20x, and, yes, even 100x speed-ups when they run (and optimize) their code on GPUs or other accelerators. This isn’t an NVIDIA-only statement; it also covers customers using IBM Cell processors and FPGAs. For highly parallelized code, accelerators run rings around CPUs – with the term ‘rings’ being defined as ‘greater than only 2.5x faster.’

So does that mean that accelerators are the ‘one true answer’ for HPC? Nope, there isn’t such a thing in this industry… but accelerators like GPUs and FPGAs can provide a huge benefit when used in the right situation – which is why they’re a hot product right now, and why Intel felt the need to slap them down. I wonder if this paper would have been written if Larabee had survived? Hmm… now that’s an interesting question…

Leave a Reply

Your email address will not be published. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <strike> <strong>