I’m trying to optimize an application that already uses Optix to determine which facets of complex 3-D models are visible. The code creates about 400,000 rays, then executes the Optix query, which takes about 0.5 seconds (Tesla K40c) on Ubuntu Linux.
Does it make sense to look at the host->GPU transfer time for the data structure that contains the 400,000 rays? Can that be improved? My guess is that it can’t, but a smart person told me to look into this. Is there anything I can do to improve that? Currently, the transfer happens automatically within the API call to execute the Optix query, so I don’t even know if it’s possible to optimize the data transfer.
I’m basically clueless about how the transfer works, so if it does not make sense to look into this, feel free to tell me.
Thanks for any advice in advance.