AI Catalyst Conference: NLP with ChatGPT (and other Large Language Models)
Published by Pearson
Transformer Architectures from Development to Deployment
- Discover the astounding state-of-the-art in Natural Language Processing (NLP) that is enabled by Large Language Models (LLMs) like ChatGPT and T5
- Understand Attention and Transformers, as well as how these essential modern NLP concepts relate to Deep Learning and LLMs
- Survey the staggeringly broad range of LLMs’ natural-language capabilities
- Learn how to use LLMs in practice, including how to train and deploy them into production NLP applications
Large Language Models (LLMs) such as GPT series architectures have dramatically accelerated the natural language processing (NLP) capabilities of machines in recent years. These capabilities, facilitated by LLMs’ hundreds of billions of model parameters, approach or exceed human-level performance on a staggeringly broad set of natural-language tasks — often without any task-specific training being required. In this event, leading subject-matter experts introduce LLMs and their associated concepts (e.g., Transformers, Attention), survey LLMs’ breadth of capabilities, and provide the best practices on how to leverage LLMs efficiently and confidently in order to supercharge your own natural-language applications.
AI Catalyst
The AI Catalyst event from Pearson brings together fresh voices in AI to make complex topics understandable and actionable. Host Jon Krohn guides the conversation and explains how to bring state-of-the-art methods into practice. Gain new information or a different perspective to make an impact in your job and in the world.
What you’ll learn and how you can apply it
- Large Language Models (LLMs)
- Attention
- Transformers
- The breadth of state-of-the-art NLP applications
And you’ll be able to:
- Select an appropriate LLM architecture for a given NLP application
- Prompt pre-trained LLMs like ChatGPT and GPT-3 to effectively produce your desired output
- Train and deploy LLMs into production NLP applications
- Potentially accelerate your data science roadmap by months or years by leveraging a pre-trained LLM instead of needing to train individual task-specific models from scratch yourself
This live event is for you because...
- You’d like to appreciate the staggering breadth of NLP and Deep Learning capabilities
- You are a data scientist, software developer, ML engineer, or other technical professional who would like to be able incorporate new NLP approaches into real-world applications
Prerequisites
- All you need is an interest in how AI can impact you and your organization.
- Feel free to bring questions for the experts.
Recommended Follow-up
- Attend: Deploying GPT and Large Language Models by Sinan Ozdemir
- Attend: Hands-on Natural Language Generation and GPT by Sinan Ozdemir
- Read: Chapter 15 of Learning Deep Learning by Dr. Magnus Ekman (book)
- Watch: NLP using Transformer Architectures by Aurélien Géron (video)
- For a more general introduction to deep learning, check out the Deep Learning: The Complete Guide playlist by Dr. Jon Krohn
Schedule
The time frames are only estimates and may vary according to how the class is progressing.
Jon Krohn: Welcome (10 minutes)
Sinan Ozdemir: Introduction to Large Language Models (30 minutes)
- We can’t talk about state-of-the-art Natural Language Processing (NLP) without talking about Transformers and large language models (LLMs) like BERT, GPT, and T5. We will explore a brief history of modern NLP up to the rise of attention-based models and Transformers including the proliferation of LLMs that continues to this day along with all of the good and sometimes the not-so-good outcomes. We will overview the major architectures that influence the tasks and models that dominate NLP while peeking under the hood to understand how LLMs learn to read, write, and do so much more.
- Sinan Ozdemir is an active lecturer focusing on large language models and a former lecturer of data science at the Johns Hopkins University. He is the author of multiple textbooks on data science and machine learning including The Principles of Data Science. Sinan is currently the founder of Shiba.art which uses AI to unlock new fine art markets and was previous founder of the recently-acquired Kylie.ai, an enterprise-grade conversational AI platform with RPA capabilities. He holds a Master’s Degree in Pure Mathematics from Johns Hopkins University and is based in San Francisco.
Jon and Sinan Discussion + Q&A (15 min)
Melanie Subbiah: The Broad Range of LLM Capabilities (30 minutes)
- Large language models have unlocked a huge number of exciting applications in the real world that were not possible before — capabilities that are creative, useful, and profitable. Through interactive demos of GPT-3, we will explore a broad range of these use cases, giving participants more intuition for how large language models have been effective.
- Melanie Subbiah is a third-year PhD student in NLP at Columbia University where she researches narrative summarization and aspects of online text safety. Before starting graduate school, she was one of the lead authors on the GPT-3 paper, building out the evaluation suite for that work and helping early customers use the OpenAI API for their projects. Prior to that, she researched autonomous systems at Apple. Melanie obtained her Bachelor’s in computer science from Williams College.
Jon and Melanie Discussion + Q&A (15 min)
- Break 10 min
Shaan Khosla: Training and Deploying LLMs (30 minutes)
- Shaan will cover practical LLM tips over the full NLP lifecycle. These include topics such as efficient training practices, validation methods, and productionization considerations to ensure your design is optimized for implementation within your real-world natural-language application.
- Shaan Khosla is a data scientist at Nebula where he researches, designs, and develops NLP models. He’s previously worked at Bank of America on an internal machine learning consulting team, where he used LLMs to build proof of concept systems for various lines of business. Shaan holds a BSBA in Computer Science and Finance from the University of Miami and is currently completing a Master’s in Data Science at NYU. He has published multiple peer-reviewed papers applying LLMs, topic modeling, and recommendation systems to the fields of biochemistry and healthcare.
Jon and Shaan Discussion + Q&A (15 min)
Jon Krohn: Closing Remarks (10 minutes)
Your Host
Jon Krohn
Jon Krohn is Co-Founder and Chief Data Scientist at the machine learning company Nebula. He authored the book Deep Learning Illustrated, an instant #1 bestseller that was translated into seven languages. He is also the host of SuperDataScience, the data science industry’s most listened-to podcast. Jon is renowned for his compelling lectures, which he offers at leading universities and conferences, as well as via his award-winning YouTube channel. He holds a PhD from Oxford and has been publishing on machine learning in prominent academic journals since 2010.