The Kaitchup – AI on a Budget

The Kaitchup – AI on a Budget

Share this post

The Kaitchup – AI on a Budget
The Kaitchup – AI on a Budget
Qwen2.5 QLoRA, LoRA, and Full Fine-tuning on Your Computer
Copy link
Facebook
Email
Notes
More

Qwen2.5 QLoRA, LoRA, and Full Fine-tuning on Your Computer

Don't wait for Llama 4 or Qwen3. Qwen2.5 is already a significant progress.

Benjamin Marie's avatar
Benjamin Marie
Sep 23, 2024
∙ Paid
12

Share this post

The Kaitchup – AI on a Budget
The Kaitchup – AI on a Budget
Qwen2.5 QLoRA, LoRA, and Full Fine-tuning on Your Computer
Copy link
Facebook
Email
Notes
More
9
1
Share

Alibaba released Qwen2.5. The models are available in many sizes: 0.5B, 1.5B, 3B, 7B, 14B, 32B, and 72B, as base and instruct models:

  • Hugging Face collection: Qwen2.5

With all these sizes, Qwen2.5 offers flexibility across various hardware configurations. The 7B and 14B models are ideal for 24 GB GPUs, and they support efficient fine-tuning with LoRA and QLoRA. For full fine-tuning, smaller models are a better fit.

The Kaitchup – AI on a Budget is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.

In this article, I explain how to fine-tune Qwen2.5, step by step. We will compare QLoRA, LoRA, and full fine-tuning speed and memory consumption. Since many model sizes are available, even if you only have a small 8 GB GPU, you will find a Qwen2.5 model that can be fine-tuned on your hardware.

The code for fine-tuning Qwen2.5 is also implemented in this notebook:

Get the notebook (#106)

If you need GPUs, I recommend RunPod (referral link). I mainly use RunPod now, especially for long training.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 The Kaitchup
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More