Hi everyone,
I was doing some speed tests on our new Tesla C2070. Let me begin by saying that for most of my actual real code the Tesla is between 1 and 3 times faster than one of our old GeForce 9800 GX2. However, for small problem sizes the GeForce is often considerably faster. This made me do some speed tests from the code examples, with the following results:
== clock ==
[clock] starting…
Using CUDA device [0]: Tesla C2070
time = 390322
[clock] test results…
PASSED
[clock] starting…
Using CUDA device [1]: GeForce 9800 GX2
time = 20950
[clock] test results…
PASSED
== eigenvalues ==
[eigenvalues] starting…
Using CUDA device [0]: Tesla C2070
Matrix size: 2048 x 2048
Precision: 0.000010
Iterations to be timed: 100
Result filename: ‘eigenvalues.dat’
Gerschgorin interval: -2.894310 / 2.923303
Average time step 1: 34.177879 ms
Average time step 2, one intervals: 11.691751 ms
Average time step 2, mult intervals: 0.006080 ms
Average time TOTAL: 45.962509 ms
[eigenvalues] test results…
PASSED
[eigenvalues] starting…
Using CUDA device [1]: GeForce 9800 GX2
Matrix size: 2048 x 2048
Precision: 0.000010
Iterations to be timed: 100
Result filename: ‘eigenvalues.dat’
Gerschgorin interval: -2.894310 / 2.923303
Average time step 1: 12.691402 ms
Average time step 2, one intervals: 3.923561 ms
Average time step 2, mult intervals: 0.003820 ms
Average time TOTAL: 16.648130 ms
[eigenvalues] test results…
PASSED
Is this normal? Should the $2500 Tesla be (much!) slower than the $500 GX2?