Back to Projects
Gen AIFine-Tuning Workshop Series - Vizuara

Fine-Tuning Series: GPT-2 to Advanced Techniques

Comprehensive 5-day fine-tuning series covering RAG vs Fine-Tuning fundamentals, instruction fine-tuning GPT-2 from scratch, scaling efficiently with Unsloth, LoRA/QLoRA, Prefix tuning, Soft-Prompts, Multimodal fine-tuning in JAX, and cutting-edge research implementations like Subliminal Learning and RAFT.

October 20, 2024
Fine-Tuning Workshop Series - Vizuara
Fine-Tuning Series: GPT-2 to Advanced Techniques

Key Learnings

Fine-tuning is about teaching AI to specialize. I learned the fundamental trade-offs: RAG = Dynamic Knowledge + Low Training Cost, while Fine-Tuning = Specialization + High Consistency. According to research on Theoretical Limitations of Embeddings, embeddings alone cannot capture all fine-grained semantic relationships - they flatten meaning into fixed dimensions. That's why Fine-Tuning becomes crucial when tasks demand deep semantic understanding beyond retrieval. I fine-tuned GPT-2 from scratch, learning the anatomy of alignment - how models internalize instruction patterns, how weights shift for behavioral tuning, and the foundations that make InstructGPT and Alpaca possible. Unsloth revolutionized efficiency with 4-bit quantization (QLoRA), LoRA adapters, and optimized kernels enabling 7B models on consumer GPUs with <10GB VRAM. The series culminated in research implementations of Subliminal Learning (behavioral transfer in LLMs) and RAFT (Retrieval-Augmented Fine-Tuning) for domain-specific RAG.

Features

1
Day 1: RAG vs Fine-Tuning - When to use what, theoretical foundations
2
Day 2: Instruction fine-tuning GPT-2 from scratch (Alpaca-style)
3
Day 3: Efficient fine-tuning with Unsloth (2-5x faster, 70% less VRAM)
4
Day 4: LoRA, QLoRA, Prefix, Soft-Prompt, and Multimodal fine-tuning
5
Day 5: Research frontiers - Subliminal Learning and RAFT implementations
6
Full training pipelines with evaluation and from-scratch implementations

Technologies Used

PythonPyTorchHuggingFaceUnslothJAXTransformerPEFTbitsandbytes

Tags

Fine-TuningGPT-2LoRAQLoRAUnslothResearchRAFTSubliminal Learning