The Kaitchup – AI on a Budget
Subscribe
Sign in
Home
Notes
AI Notebooks
The Kaitchup's Book
Weekly Kaitchup
Tutorials
The Kaitchup Index
Archive
About
Tutorials
Latest
Top
Discussions
DenseMixer: Smarter MoE Routing That Doesn’t Break LoRA and QLoRA
Better MoE training for a slightly higher cost
Sep 8
•
Benjamin Marie
5
1
Gemma 3 270M: Can Tiny Models Learn New Tasks?
A case study with machine translation
Sep 1
•
Benjamin Marie
15
10
NVFP4: Same Accuracy with 2.3x Higher Throughput for 4-Bit LLMs
How to quantize LLMs with NVFP4
Aug 25
•
Benjamin Marie
9
6
How to Run Unsloth on Multi-GPU Setups: Data-Parallel or Model-Parallel
Step-by-step fixes for running Unsloth across GPUs
Aug 11
•
Benjamin Marie
6
3
Gemma 3n: Fine-Tuning, Inference, and Submodel Extraction
Running Gemma 3n with vLLM and fine-tuning with TRL
Jun 30
•
Benjamin Marie
10
RAG with Qwen3 Embedding and Qwen3 Reranker
How to use embedding and reranker models to efficiently retrieve only the most relevant chunks or documents given a user query
Jun 19
•
Benjamin Marie
24
4
RTX 6000 Pro vs H100 & A100: Best Single-GPU Choice for Fast, Low-Cost LLM Fine-Tuning
Faster, cheaper single-GPU training
Jun 16
•
Benjamin Marie
14
1
Fine-Tuning 2-Bit Qwen3 Models on Your Computer
Code and best practices
Jun 9
•
Benjamin Marie
12
Qwulu 3: Fine-Tuning Qwen3 Base with LoRA and TULU 3's Supervised Fine-Tuning Recipe
Can a supervised fine-tuning recipe that works effectively on Llama 3.1 be applied directly to Qwen3?
Jun 5
•
Benjamin Marie
7
Boost 2-Bit LLM Accuracy with EoRA
A training-free solution for extreme LLM compression
May 19
•
Benjamin Marie
6
LoRA at Scale on a Consumer GPU: Does It Work?
Reproducing TULU 3 SFT on Consumer Hardware Using LoRA and Unsloth
May 12
•
Benjamin Marie
7
3
Fine-Tuning Qwen3: Base vs. Reasoning Models
Is it reasonable to fine-tune a "reasoning" model?
May 8
•
Benjamin Marie
11
2
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts