Practical guide: Fine-tuning Qwen3 with LoRA
Details
Ivan Potapov - Research Engineer, Zalando SE
In this workshop, we fine-tune Qwen models with parameter-efficient adapters using two complementary approaches: Soft Prompt token tuning and LoRA SFT, with an optional KL-anchored SFT term to keep the model’s behavior close to the base while adding new styles and formats. You’ll see how to prepare open-source data (Dolly 15k), render with chat templates, run short training loops, and monitor validation loss/perplexity with stepwise evaluations.
A tiny KL toy example explains per‑token contributions to H(P), H(P,Q), and KL(P||Q), making the “anchoring” intuition concrete. By the end, you’ll know how to apply Soft Prompt for quick style steering, LoRA for deeper adaptation, and KL regularization to reduce drift and forgetting—plus how to save/load LoRA adapters for deployment.
About the speaker:
Ivan Potapov is a Research Engineer at Zalando, specializing in search. He has taught workshops on data engineering, AI agents, and LLM alignment, helping practitioners bridge software engineering with applied machine learning.
Join our slack: https://datatalks.club/slack.html
