Hands-On Large Language Models
I am thrilled to introduce the “Hands-On Large Language Models” book I wrote with Jay Alammar!
With the incredible pace of LLM development, learning about these techniques can be overwhelming. Throughout this book, we take an intuition-first approach through visual storytelling with almost 300 custom-made images in the final release.
This book is for those interested in this exciting field. Whether you are a beginner or more advanced, we believe there is something to be found for everyone!
All code is freely available on GitHub making it easy for you to get started with the inner workings of LLMs.
Visual Storytelling
Throughout our careers, Jay and I have created educative content in AI with a large focus on illustrations. With the complex and often mathematics-heavy nature of AI, the field quickly becomes overwhelming. By leveraging a visual style of diving into these techniques, we aim to focus on understanding these techniques moreso than writing down the equations that make them up.
Great examples of our visual styles can be found here:
The Illustrated Transformer - Jay Alammar
A Visual Guide to Mamba and State Space Models - Maarten Grootendorst
If these styles appeal to you, then you will definitely like this book!
Table of Contents
Here, you will find an overview of the chapters that will be in the book:
Part 1 - Concepts
Introduction to Language Models
Token Embeddings
Looking Inside Transformer LLMs
Part 2 - Using Pre-Trained Language Models
Text Classification
Text Clustering and Topic Modeling
Prompt Engineering
Advanced Text Generation Techniques and Tools
Semantic Search and Retrieval Augmented Generation
Multimodal Large Language Models
Part 3 - Training and Fine-Tuning
Creating Text Embedding Models
Fine-Tuning Representation Models for Classification
Fine-Tuning Generation Models