Minimal example training a Qwen3.5 0.8B lora on AMD RYZEN AI MAX+ 395 APU
Find a file
2026-03-26 20:50:38 +11:00
.envrc init 2026-03-26 20:40:59 +11:00
.gitignore init 2026-03-26 20:40:59 +11:00
README.md add readme 2026-03-26 20:50:38 +11:00
shell.nix init 2026-03-26 20:40:59 +11:00
train.py init 2026-03-26 20:40:59 +11:00

LoRA Training on AMD

Overview

shell.nix

Provides a development environment compatible within Nix. It handles:

  • Installing PyTorch optimized for AMD ROCm (rocm7.2).
  • Installing unsloth and unsloth-zoo for efficient fine-tuning.
  • Installing marimo and ipython as QOL.

train.py

A marimo script that executes the fine-tuning process:

  • Loads the unsloth/Qwen3.5-0.8B model in 16-bit.
  • Prepares a sample dataset (laion/OIG/unified_chip2.jsonl).
  • Configures Parameter-Efficient Fine-Tuning (PEFT) using LoRA (Rank 16).
  • Sets up an SFTTrainer (using 8-bit AdamW) and trains the model for 100 steps, saving results to outputs_qwen35.