I have downloaded Mistral-7B-Instruct-v0.2 model from huggingface and want to convert that model to Triton Inference Server supported framework and Run the model using Triton Inference server. I need support doc which provides steps to deploy and also infrastructure details.
Hi @jutursundarkumar.reddy please see our docs here Deploying Hugging Face Transformer Models in Triton — NVIDIA Triton Inference Server.