Hi @esteban.gallardo, it looks like you are trying to load a GPTQ model with the llama.cpp loader - llama.cpp only works with GGUF models. IIRC, text-generation-webui expects the .gguf model file to be saved under it’s model directory (/data/models/text-generation-webui)
By the way, for llama.cpp loader you should increase the n-gpu-layers setting (typically to the max), otherwise it will not use GPU.