I am using TensorRT C++ API, and I want to avoid data copy using for loops to store the image pixels to TensorRT Buffer.
If I know the CPU pointer of the lowest address pixel (i.e., the first pixel) of the image, and I know the image is stored
in a contiguous memory block, is there a way to wrap the data with TensorRT buffer directly?
I have the same question for GPU pointer. I want to wrap image data stored on GPU with a TensorRT directly without any data copy given that I have the GPU pointer to the first pixel.
libtorch has a function doing this called: torch::from_blob. Is there a similar function in TensorRT?
TensorRT Version: TensorRT-22.214.171.124.Windows10.x86_64.cuda-10.2.cudnn7.6
GPU Type: GPU 1NVIDIA GeForce GTX 1660 Ti with Max-Q
Nvidia Driver Version:
CUDA Version: 10.2
CUDNN Version: 7.6
Operating System + Version: Windows 10
Python Version (if applicable):
TensorFlow Version (if applicable):
PyTorch Version (if applicable):
Baremetal or Container (if container which image + tag):
Please attach or include links to any models, data, files, or scripts necessary to reproduce your issue. (Github repo, Google Drive, Dropbox, etc.)
Steps To Reproduce
- Exact steps/commands to build your repro
- Exact steps/commands to run your repro
- Full traceback of errors encountered