Hi everyone,

I was doing some speed tests on our new Tesla C2070. Let me begin by saying that for most of my actual real code the Tesla is between 1 and 3 times faster than one of our old GeForce 9800 GX2. However, for small problem sizes the GeForce is often considerably faster. This made me do some speed tests from the code examples, with the following results:

== clock ==

[clock] starting…

Using CUDA device [0]: Tesla C2070

time = 390322

[clock] test results…

PASSED

[clock] starting…

Using CUDA device [1]: GeForce 9800 GX2

time = 20950

[clock] test results…

PASSED

== eigenvalues ==

[eigenvalues] starting…

Using CUDA device [0]: Tesla C2070

Matrix size: 2048 x 2048

Precision: 0.000010

Iterations to be timed: 100

Result filename: ‘eigenvalues.dat’

Gerschgorin interval: -2.894310 / 2.923303

Average time step 1: 34.177879 ms

Average time step 2, one intervals: 11.691751 ms

Average time step 2, mult intervals: 0.006080 ms

Average time TOTAL: 45.962509 ms

[eigenvalues] test results…

PASSED

[eigenvalues] starting…

Using CUDA device [1]: GeForce 9800 GX2

Matrix size: 2048 x 2048

Precision: 0.000010

Iterations to be timed: 100

Result filename: ‘eigenvalues.dat’

Gerschgorin interval: -2.894310 / 2.923303

Average time step 1: 12.691402 ms

Average time step 2, one intervals: 3.923561 ms

Average time step 2, mult intervals: 0.003820 ms

Average time TOTAL: 16.648130 ms

[eigenvalues] test results…

PASSED

Is this normal? Should the $2500 Tesla be (much!) slower than the $500 GX2?