Book description
Since their introduction in 2017, transformers have quickly become the dominant architecture for achieving state-of-the-art results on a variety of natural language processing tasks. If you're a data scientist or coder, this practical book -now revised in full color- shows you how to train and scale these large models using Hugging Face Transformers, a Python-based deep learning library.
Transformers have been used to write realistic news stories, improve Google Search queries, and even create chatbots that tell corny jokes. In this guide, authors Lewis Tunstall, Leandro von Werra, and Thomas Wolf, among the creators of Hugging Face Transformers, use a hands-on approach to teach you how transformers work and how to integrate them in your applications. You'll quickly learn a variety of tasks they can help you solve.
- Build, debug, and optimize transformer models for core NLP tasks, such as text classification, named entity recognition, and question answering
- Learn how transformers can be used for cross-lingual transfer learning
- Apply transformers in real-world scenarios where labeled data is scarce
- Make transformer models efficient for deployment using techniques such as distillation, pruning, and quantization
- Train transformers from scratch and learn how to scale to multiple GPUs and distributed environments
Publisher resources
Table of contents
- Foreword
- Preface
- 1. Hello Transformers
- 2. Text Classification
- 3. Transformer Anatomy
-
4. Multilingual Named Entity Recognition
- The Dataset
- Multilingual Transformers
- A Closer Look at Tokenization
- Transformers for Named Entity Recognition
- The Anatomy of the Transformers Model Class
- Tokenizing Texts for NER
- Performance Measures
- Fine-Tuning XLM-RoBERTa
- Error Analysis
- Cross-Lingual Transfer
- Interacting with Model Widgets
- Conclusion
- 5. Text Generation
- 6. Summarization
- 7. Question Answering
- 8. Making Transformers Efficient in Production
- 9. Dealing with Few to No Labels
- 10. Training Transformers from Scratch
- 11. Future Directions
- Index
- About the Authors
Product information
- Title: Natural Language Processing with Transformers, Revised Edition
- Author(s):
- Release date: May 2022
- Publisher(s): O'Reilly Media, Inc.
- ISBN: 9781098136796
You might also like
book
Generative Deep Learning, 2nd Edition
Generative AI is the hottest topic in tech. This practical book teaches machine learning engineers and …
book
Deep Learning for Coders with fastai and PyTorch
Deep learning is often viewed as the exclusive domain of math PhDs and big tech companies. …
book
Transformers for Natural Language Processing - Second Edition
OpenAI's GPT-3, ChatGPT, GPT-4 and Hugging Face transformers for language tasks in one book. Get a …
book
Hands-On Generative AI with Transformers and Diffusion Models
Learn to use generative AI techniques to create novel text, images, audio, and even music with …