Hello everyone,
Has anyone tried the tutorial from Unsloth https://unsloth.ai/docs/basics/fine-tuning-llms-with-nvidia-dgx-spark-and-unsloth ?
I’m facing some issues with building the Dockerfile on my machine (building xformers from source fails). However, I managed to run it using this setup script:
uv venv
source .venv/bin/activate
uv pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu130
uv pip install transformers peft datasets
uv pip install --no-deps unsloth unsloth-zoo
uv pip install --no-deps bitsandbytes
uv pip install --upgrade torchao
uv pip install --upgrade unsloth unsloth-zoo transformers
uv pip install xformers
The article mentions a training time of around 4 hours, but on my machine, I was only able to reduce it from 72 hours to approximately 24 hours by tweaking generation_length and batch size. Still, it’s quite far from the 4-hour target.
Has anyone else tried this tutorial? What kind of performance or training times are you getting?
Thanks in advance!