The Kaitchup – AI on a Budget
Subscribe
Sign in
Home
Notes
AI Notebooks
The Kaitchup Pro
The Kaitchup's Book
AI Toolboxes
Tutorials
Models
LLM Leaderboards
Archive
About
Tutorials
Latest
Top
Discussions
LLM Alignment: Searching for Optimal ORPO Hyperparameters
Higher learning rate and beta
6 hrs ago
•
Benjamin Marie
4
Share this post
The Kaitchup – AI on a Budget
LLM Alignment: Searching for Optimal ORPO Hyperparameters
Copy link
Facebook
Email
Notes
More
The Recipe for Extremely Accurate and Cheap Quantization of 70B+ LLMs
Cost and accuracy for quantizing large models to 4-bit and 2-bit
Nov 25
•
Benjamin Marie
7
Share this post
The Kaitchup – AI on a Budget
The Recipe for Extremely Accurate and Cheap Quantization of 70B+ LLMs
Copy link
Facebook
Email
Notes
More
DPO Full Training vs. LoRA: How Good is LoRA for DPO Training?
One model, two adapters
Nov 18
•
Benjamin Marie
6
Share this post
The Kaitchup – AI on a Budget
DPO Full Training vs. LoRA: How Good is LoRA for DPO Training?
Copy link
Facebook
Email
Notes
More
Torch Compile: 2x Faster Llama 3.2 with Low Effort
But it will depend on your GPU
Nov 11
•
Benjamin Marie
5
Share this post
The Kaitchup – AI on a Budget
Torch Compile: 2x Faster Llama 3.2 with Low Effort
Copy link
Facebook
Email
Notes
More
8
LLM as a Judge: Evaluate Your LLMs with Another LLM
A good evaluation framework for quick feedback and monitoring
Nov 7
•
Benjamin Marie
11
Share this post
The Kaitchup – AI on a Budget
LLM as a Judge: Evaluate Your LLMs with Another LLM
Copy link
Facebook
Email
Notes
More
Llama 3.2 Embeddings: Training and Evaluation with LLM2Vec
A step-by-step tutorial
Nov 4
•
Benjamin Marie
11
Share this post
The Kaitchup – AI on a Budget
Llama 3.2 Embeddings: Training and Evaluation with LLM2Vec
Copy link
Facebook
Email
Notes
More
2
The Impact of the Calibration Dataset for AutoRound and AWQ Quantization
Should you choose the calibration dataset?
Oct 31
•
Benjamin Marie
4
Share this post
The Kaitchup – AI on a Budget
The Impact of the Calibration Dataset for AutoRound and AWQ Quantization
Copy link
Facebook
Email
Notes
More
3
bitnet.cpp: Efficient Inference with 1-Bit LLMs on your CPU
How to run "1-bit" (but 1.58-bit) LLMs made of ternary weights packed to 2-bit
Oct 28
•
Benjamin Marie
11
Share this post
The Kaitchup – AI on a Budget
bitnet.cpp: Efficient Inference with 1-Bit LLMs on your CPU
Copy link
Facebook
Email
Notes
More
3
Generate Videos on Your Computer with Pyramid Flow
AI-generated videos with 768p resolution and 24 FPS
Oct 24
•
Benjamin Marie
7
Share this post
The Kaitchup – AI on a Budget
Generate Videos on Your Computer with Pyramid Flow
Copy link
Facebook
Email
Notes
More
Train and Serve an AI Chatbot Based on Llama 3.2
Efficient supervised fine-tuning with TRL
Oct 17
•
Benjamin Marie
9
Share this post
The Kaitchup – AI on a Budget
Train and Serve an AI Chatbot Based on Llama 3.2
Copy link
Facebook
Email
Notes
More
8
Fast Speculative Decoding with Llama 3.2 and vLLM
Boost LLM inference speed with speculative decoding!
Oct 14
•
Benjamin Marie
11
Share this post
The Kaitchup – AI on a Budget
Fast Speculative Decoding with Llama 3.2 and vLLM
Copy link
Facebook
Email
Notes
More
2
Generate Synthetic Data from Personas to Train AI Chatbots
Using Personas and Efficient Inference to Create Targeted Training Data for AI Chatbots
Oct 10
•
Benjamin Marie
7
Share this post
The Kaitchup – AI on a Budget
Generate Synthetic Data from Personas to Train AI Chatbots
Copy link
Facebook
Email
Notes
More
7
Share
Copy link
Facebook
Email
Notes
More
This site requires JavaScript to run correctly. Please
turn on JavaScript
or unblock scripts