Hands-On Large Language Models

I am thrilled to introduce the “Hands-On Large Language Models” book I wrote with Jay Alammar!

With the incredible pace of LLM development, learning about these techniques can be overwhelming. Throughout this book, we take an intuition-first approach through visual storytelling with almost 300 custom-made images in the final release.

This book is for those interested in this exciting field. Whether you are a beginner or more advanced, we believe there is something to be found for everyone!

All code is freely available on GitHub making it easy for you to get started with the inner workings of LLMs.

Visual Storytelling

Throughout our careers, Jay and I have created educative content in AI with a large focus on illustrations. With the complex and often mathematics-heavy nature of AI, the field quickly becomes overwhelming. By leveraging a visual style of diving into these techniques, we aim to focus on understanding these techniques moreso than writing down the equations that make them up.

Great examples of our visual styles can be found here:

If these styles appeal to you, then you will definitely like this book!

Table of Contents

Here, you will find an overview of the chapters that will be in the book:

Part 1 - Concepts

  1. Introduction to Language Models

  2. Token Embeddings

  3. Looking Inside Transformer LLMs

Part 2 - Using Pre-Trained Language Models

  1. Text Classification

  2. Text Clustering and Topic Modeling

  3. Prompt Engineering

  4. Advanced Text Generation Techniques and Tools

  5. Semantic Search and Retrieval Augmented Generation

  6. Multimodal Large Language Models

Part 3 - Training and Fine-Tuning

  1. Creating Text Embedding Models

  2. Fine-Tuning Representation Models for Classification

  3. Fine-Tuning Generation Models