Dear OptiX Development Team,
I am writing to inquire about the planned support for the TF32 (TensorFloat32) data type within the next release of OptiX in the neural rendering routines (specifically, optixCoopVecMatMul and related functions). Currently, I observe that despite the availability of TF32 hardware support (Tensor Cores) and its ideal characteristics for deep learning - offering FP32 range with FP16-equivalent precision - the following input/output combinations are not natively supported in OptiX:
| inputType | inputInterpretation | matrixElementType | biasElementType | outputType |
|---|---|---|---|---|
| FP32 | TF32 | FP32 | FP32 | FP32 |
The current reliance primarily on FLOAT16 is often insufficient for differential neural rendering (DNR) and inverse problems, where the narrow dynamic range of FP16 frequently leads to gradient underflow or unstable numerical results.
Are there plans to introduce native TF32 support for the input vector, bias, and matrix elements in the next release of OptiX? This feature would significantly enhance the numerical stability and performance of high-fidelity neural rendering applications.
Thank you for your consideration.