Minimal example training a Qwen3.5 0.8B lora on AMD RYZEN AI MAX+ 395 APU
| .envrc | ||
| .gitignore | ||
| README.md | ||
| shell.nix | ||
| train.py | ||
LoRA Training on AMD
Overview
shell.nix
Provides a development environment compatible within Nix. It handles:
- Installing PyTorch optimized for AMD ROCm (
rocm7.2). - Installing
unslothandunsloth-zoofor efficient fine-tuning. - Installing
marimoandipythonas QOL.
train.py
A marimo script that executes the fine-tuning process:
- Loads the
unsloth/Qwen3.5-0.8Bmodel in 16-bit. - Prepares a sample dataset (
laion/OIG/unified_chip2.jsonl). - Configures Parameter-Efficient Fine-Tuning (PEFT) using LoRA (Rank 16).
- Sets up an
SFTTrainer(using 8-bit AdamW) and trains the model for 100 steps, saving results tooutputs_qwen35.