Hi all!
I would like to use Ollama along with nemotron-mini.
I decided to work the nemotron-mini:4b-instruct-q4_K_M model that is pulled from Ollama. I can see that sometimes it has issues with tool handling, especially when multiple tools are defined. It wasn’t the case with other models, that I have tried. It seems that the output JSON misses an arguments part and this is the reason Ollama doesn’t handle the tool calling properly.
I’m also using ollama-python using following example as a base solution: Ollama Python library 0.4 with function calling improvements · Ollama Blog. Due to memory constraints (8GB Orin Nano), I can’t have larger model.
Could you please guide me with the required/recommended steps to proceed?
Thank you!