Book description
AI has acquired startling new language capabilities in just the past few years. Driven by rapid advances in deep learning, language AI systems are able to write and understand text better than ever before. This trend is enabling new features, products, and entire industries. Through this book's visually educational nature, readers will learn practical tools and concepts they need to use these capabilities today.
You'll understand how to use pretrained large language models for use cases like copywriting and summarization; create semantic search systems that go beyond keyword matching; and use existing libraries and pretrained models for text classification, search, and clusterings.
This book also helps you:
- Understand the architecture of Transformer language models that excel at text generation and representation
- Build advanced LLM pipelines to cluster text documents and explore the topics they cover
- Build semantic search engines that go beyond keyword search, using methods like dense retrieval and rerankers
- Explore how generative models can be used, from prompt engineering all the way to retrieval-augmented generation
- Gain a deeper understanding of how to train LLMs and optimize them for specific applications using generative model fine-tuning, contrastive fine-tuning, and in-context learning
Publisher resources
Table of contents
- Preface
- I. Understanding Language Models
-
1. An Introduction to Large Language Models
- What Is Language AI?
- A Recent History of Language AI
- The Moving Definition of a “Large Language Model”
- The Training Paradigm of Large Language Models
- Large Language Model Applications: What Makes Them So Useful?
- Responsible LLM Development and Usage
- Limited Resources Are All You Need
- Interfacing with Large Language Models
- Generating Your First Text
- Summary
- 2. Tokens and Embeddings
- 3. Looking Inside Large Language Models
- II. Using Pretrained Language Models
- 4. Text Classification
- 5. Text Clustering and Topic Modeling
- 6. Prompt Engineering
- 7. Advanced Text Generation Techniques and Tools
- 8. Semantic Search and Retrieval-Augmented Generation
- 9. Multimodal Large Language Models
- III. Training and Fine-Tuning Language Models
- 10. Creating Text Embedding Models
- 11. Fine-Tuning Representation Models for Classification
-
12. Fine-Tuning Generation Models
- The Three LLM Training Steps: Pretraining, Supervised Fine-Tuning, and Preference Tuning
- Supervised Fine-Tuning (SFT)
- Instruction Tuning with QLoRA
- Evaluating Generative Models
- Preference-Tuning / Alignment / RLHF
- Automating Preference Evaluation Using Reward Models
- Preference Tuning with DPO
- Summary
- Afterword
- Index
- About the Authors
Product information
- Title: Hands-On Large Language Models
- Author(s):
- Release date: September 2024
- Publisher(s): O'Reilly Media, Inc.
- ISBN: 9781098150969
You might also like
book
Modern Generative AI with ChatGPT and OpenAI Models
Harness the power of AI with innovative, real-world applications, and unprecedented productivity boosts, powered by the …
book
Quick Start Guide to Large Language Models: Strategies and Best Practices for Using ChatGPT and Other LLMs
The advancement of Large Language Models (LLMs) has revolutionized the field of Natural Language Processing in …
book
Deep Learning for Coders with fastai and PyTorch
Deep learning is often viewed as the exclusive domain of math PhDs and big tech companies. …
book
Fluent Python, 2nd Edition
Don't waste time bending Python to fit patterns you've learned in other languages. Python's simplicity lets …