Can someone tell me how to benchmark LLama_v2_7b model on jetson Orin AGX with different quantization methods and also how to get perplexity score
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
How to benchmark AI processes on Jetson orin NX 8gb | 3 | 793 | October 6, 2023 | |
Running LLMs with TensorRT-LLM on Nvidia Jetson AGX Orin Dev Kit | 1 | 465 | December 8, 2024 | |
Running LMdeploy inference engine on the NVIDIA Jetson AGX Orin Devkit | 2 | 101 | January 14, 2025 | |
LLM Performance Benchmarking: Measuring NVIDIA NIM Performance with GenAI-Perf | 1 | 5 | May 6, 2025 | |
CUDA benchmark | 2 | 1351 | March 20, 2023 | |
LLMs token/sec | 2 | 892 | April 8, 2024 | |
Model Performance Request | 3 | 286 | February 20, 2024 | |
Problem: slow LLM inference speed on Jetson AGX Orin 64GB | 2 | 110 | April 8, 2025 | |
GPU benchmark test | 2 | 911 | September 20, 2023 | |
Jetson AGX Orin Inference Performance | 2 | 750 | May 19, 2022 |