Hello all,
I’m trying to use RT cores for hardware accelerated tree traversals on the GPU. To give a little more context, I’m trying to use RT cores for general purpose computing insteading of rendering graphics where I use RT cores to accelerate tree traversals where tree is the search space I need to prune to find nearest neighbors for a query point. How do I profile the RT core usage and understand the limitations of my code.