|
Real-Time Inference on Thor & RTX Pi0.5/GR00T N1.6/1.7 Thor 23 Hz RTX 5090 50-80Hz
|
|
2
|
50
|
May 4, 2026
|
|
Jetson Thor - INT8 quantization show no performance gain over FP16 (2)
|
|
5
|
294
|
March 27, 2026
|
|
Nvfp4 Dynamic Quantizer Very Slow with Bias
|
|
3
|
84
|
January 29, 2026
|
|
Jetson Thor - INT8 quantization show no performance gain over FP16
|
|
8
|
449
|
February 9, 2026
|
|
Quantized GeMM using fp32 for Q/DQ layers
|
|
0
|
59
|
January 2, 2026
|
|
Worse performance after quantization on TensorRT
|
|
1
|
211
|
December 23, 2025
|
|
Failing to load pruned model yolov7 using tensorrt model opt
|
|
1
|
61
|
November 26, 2025
|
|
TensorRT Quantization for Jetson Inference
|
|
4
|
433
|
October 17, 2025
|
|
ConvNeXT inference with int8 quantization slower on tensorRT than fp32/fp16
|
|
2
|
317
|
September 19, 2025
|
|
30% slowdown on ResNet50 with ModelOptimizer INT8 quantization (RTX 4090)
|
|
0
|
122
|
September 19, 2025
|
|
Holistically-Nested Edge Detection using TensoRT
|
|
8
|
360
|
April 9, 2025
|
|
Errors with training flux of sparsity with accelerate
|
|
2
|
114
|
March 4, 2025
|
|
TensorRT examples
|
|
1
|
114
|
February 28, 2025
|
|
How to use same tensor rt version of Jetson orin nano in desktop PC environment
|
|
4
|
170
|
March 26, 2025
|
|
INT8 Calibration with DS 6.3 worse than with DS 6.0
|
|
20
|
411
|
March 10, 2025
|
|
Is there a plan to support DLA on the next TensorRT version?
|
|
5
|
376
|
December 31, 2024
|
|
[TRT] jetson agx orion error - CaffeParser: Could not open file device GPU, failed to load networks/Googlenet/bvlc_googlenet.caffemodel
|
|
4
|
139
|
October 18, 2024
|
|
Improving the speed for fp32 for yolov10x inference from Ultralytics on Jetson AGX Orin 64g devkit
|
|
5
|
224
|
September 18, 2024
|
|
Converting an ONNX model to TensorRT Engine on a x86/64 PC and then using it on a Jetson
|
|
2
|
175
|
August 3, 2024
|
|
[New] Discord channel for triton-inference-server, tensorrt, tensorrt-llm, model-optimization
|
|
0
|
235
|
July 16, 2024
|
|
TensorRT 10.2 is not using FP8 convolution tactics when building a FP8 quantized conv model
|
|
2
|
345
|
July 10, 2024
|
|
GPUs hang when executing NIM docker container on a 4xA100
|
|
2
|
201
|
June 29, 2024
|
|
/TopK_5: K exceeds the maximum value allowed (3840)
|
|
0
|
544
|
June 11, 2024
|