Finetuning

The process of adapting a pre-trained large language model (LLM) to a specific task or dataset by continuing its training on a smaller, specialized dataset.

Parameter-Efficient Fine-Tuning (PEFT)

Given the input x 
Output h = W0x + ∆W x = W0x + BAx

Given the input x  Output h = W0x + ∆W x = W0x + BAx

image.png

image.png

Model Quantization

Model Quantization

Unsloth Fine-Tuning

How is Unsloth Faster?

Technique Benefit
Manual autograd & optimized matrix multiplication chaining Reduces unnecessary compute overhead
All performance-critical kernels rewritten in OpenAI Triton Faster training and reduced GPU memory usage

Hands-On Time!