Hello, I would like to know if NIM can be used to deploy fine tuning or custom models. I deployed the MIXTRAL-8x7b-INSTRUCT-2-A100.24.01 model through NIM following this tutorial and now I’m trying to launch the optimized mixtral engines (checkpoints from HF) with tensorRT-LLM (ensemble, postprocessing, preprocessing, tensorrt_llm and tensorrt_llm_bls) but I have not been able to, Is there a tutorial on how to launch finetunning/custom models through NIM? Is there documentation of the structure of the model repository that NIM needs to deploy a model?
yuqih
2
Hi, I think this tutorial is tailored for you: Parameter-Efficient Fine-Tuning - NVIDIA Docs
It introduces how to obtain LoRA adapters from HF and then serve with NIM.
Hope it helps!
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
Deploying Fine-Tuned AI Models with NVIDIA NIM | 1 | 119 | November 26, 2024 | |
Container image (nim) construction guide for models where a nim doesn't exist? | 2 | 129 | November 29, 2024 | |
NVIDIA NIM으로 파인 튜닝된 AI 모델 배포하기 | 1 | 9 | December 2, 2024 | |
I have an issue when installing NVIDIA NIM on laptop | 7 | 582 | May 21, 2025 | |
Misral 및 Mixral 모델을 위한 새로운 NVIDIA NIM을 통해 AI 프로젝트 지원 | 1 | 4 | July 26, 2024 | |
NIM custom inference engine | 1 | 29 | April 1, 2025 | |
Is it currently possible to deploy our own models on NVIDIA's cloud and use NIM for inference? | 2 | 175 | July 24, 2024 | |
Customizing NVIDIA NIMs for Domain-Specific Needs with NVIDIA NeMo | 1 | 61 | July 10, 2024 | |
Customize Generative AI Models for Enterprise Applications with Llama 3.1 | 2 | 49 | July 25, 2024 | |
Access to NVIDIA NIM Now Available Free to Developer Program Members | 8 | 387 | February 26, 2025 |