I’m looking for the best hardware setup for [mostly double float] number crunching. The gtx titan and the hd7970 are the candidates. I prefer the nvidia drivers, tools and the titan has 6gb memory, but it’s mighty expensive and the 7970 seems to have it covered in most benchmarks. However, dgemm performance is very important to my application and from what I gathered the 7970 is capable of only ~600 gflops while the titan is at about two times as much. I wonder whether that matches your experience. Maybe the peak performance figures are misleading and one has to see the whole curve. I also wonder how much of the difference originates from the libraries (cublas/clblas).
I’m sorry if the above is a bit incoherent, any insight into double float [dgemm heavy or not] workloads is appreciated.