I’m new to OptiX, and I try to understand some underlying mechanisms to improve the performances.
I make some tests to know the impact of the launch size (ie: width and height). I make those value vary, like, width=height, width=1 or height=1, for the same total size. For small values it seems there is no noticeable differences for the computational time.
But for higher total size I got some problem. For a size of 40 000, I have the same computational time for width=height, and height=1. But when I set width=1, it does not work, I got this error:
OptiX error: Unknown error (Details: Function "RTresult _rtContextLaunch2D(RTcontext, unsigned int, RTsize, RTsize)" caught exception: Encountered a CUDA error: cudaDriver().CuEventSynchronize( m_event ) returned (700): Illegal address)
The documentation says the product of width and depth must be smaller than 2^32. But what about the height?
And in my case, the total size is 40000, which is far smaller than 2^32!!!
Is the maximum size depends of the computations we are doing ?
In any case, how can we know the maximum size, to add some safe guards to avoid this error?
For information:
OS: Linux CentOS7
GC: GeForce RTX 2080 TI
Cuda: 10.1
OptiX: 6
Drivers: 418.67
Thanks