Snowflake Fundamentals in 3 Weeks
Published by O'Reilly Media, Inc.
Processing and transforming data at scale
In this course you’ll:
- Learn how to architect and administer Snowflake solutions
- Understand how to build effective data pipelines with Snowflake
- Explore processing and sharing data with Snowflake
Snowflake is a modern, cloud-based data platform that has gained widespread attention and adoption in the data industry.
Join expert Tomas Sobotik to learn the fundamentals and best practices for working on the Snowflake platform, including how to properly design Snowflake infrastructure, how to process, manage, and share data, and how to tune queries in order to achieve better performance and lower costs.
Week 1: Learning Snowflake Architecture and Fundamental Concepts
In week 1, you’ll learn the basics of Snowflake architecture, data loading principles, and how to work with Snowflake compute clusters.
Week 2: Working with Snowflake Data Pipelines, Caching, and Storage
This week, you’ll learn about caching, storage, governance, and features related to building an effective data pipeline.
Week 3: Data Sharing and Performance Tuning in Snowflake
In the final week, you’ll learn techniques for performance tuning, how to share your data, and how to effectively utilize dashboarding and serverless features.
NOTE: With today’s registration, you’ll be signed up for all three weeks. Although you can attend any of the sessions individually, we recommend participating in all three weeks.
What you’ll learn and how you can apply it
- Load and share data
- Understand and use caches
- Optimize storage
This live event is for you because...
- You're a data architect who wants to design cloud-based data solutions.
- You're a data engineer who’s looking to use best practices for data loading and performance tuning on the platform.
- You're a data analyst who needs to know how to effectively prepare and share data securely in real time.
Prerequisites
- Sign up for a Snowflake trial account just before the course starts to have it available for all weeks (this is optional if you would like participate in the hands-on exercises)
- Attend SQL Fundamentals for Data (live online training with Thomas Nield)
- Explore Introducing SQL and Relational Databases (video course)
- Prepare your local Python environment for running Snowpark API. Guide is attached here.
Recommended preparation:
- Take SQL Fundamentals for Data (live online training course with Thomas Nield)
- Explore Introducing SQL and Relational Databases (video course)
Recommended follow-up:
- Read Snowflake: The Definitive Guide (book)
Schedule
The time frames are only estimates and may vary according to how the class is progressing.
Week 1: Learning Snowflake Architecture and Fundamental Concepts
Snowflake architecture (20 minutes)
- Presentation: Snowflake architecture layers
- Group discussion: Snowflake architecture knowledge check
- Q&A
Snowflake key concepts (20 minutes)
- Presentation: Snowflake key concepts: supported platforms, scalability, separation of workloads, and more
- Group discussion: Snowflake concepts knowledge check
- Q&A
Virtual warehouses (30 minutes)
- Presentation: Understanding virtual warehouses
- Hands-on exercise: Create your virtual warehouse
- Q&A
- Break
Billing (20 minutes)
- Presentation: How billing works in Snowflake
- Group discussion: Snowflake billing knowledge check
- Q&A
Snowflake UI walk-through (35 minutes)
- Presentation and demo: Basics features of the Snowsight tool; how to work with the UI live
- Hands-on exercise: Create basic database objects
- Q&A
- Break
Data-loading principles (30 minutes)
- Presentation: Introduction of database objects for data loading (stages, file formats, etc.)
- Hands-on exercise: Create your file format and stage
- Q&A
Copy command (25 minutes)
- Presentation: How to load data into Snowflake
- Hands-on exercise: Load your dataset into Snowflake
- Q&A
Week 2: Working with Snowflake Data Pipelines, Caching, and Storage
Snowflake caches (30 minutes)
- Presentation: Understanding all types of Snowflake caches
- Hands-on exercise: Explore using caches
- Q&A
Snowflake security in a nutshell (30 minutes)
- Presentation: Security principles
- Hands-on exercise: Set up the RBAC model
- Q&A
- Break
Storage (25 minutes)
- Presentation: Snowflake table types; time travel; cloning
- Hands-on exercise: Create different tables; clone data
- Q&A
Semistructured data (30 minutes)
- Presentation: Semistructured data in Snowflake
- Hands-on exercise: Flatten a JSON
- Q&A
- Break
Data governance (25 minutes)
- Presentation: Data governance features walk-through (data masking, tags, row access policies)
- Hands-on exercise: Make governance-related objects
- Q&A
Streams and tasks (40 minutes)
- Presentation: How to use streams and tasks for data pipelines
- Hands-on exercise: Create your pipeline with streams and tasks
- Q&A
Week 3: Data Sharing and Performance Tuning in Snowflake
Data Sharing (25 minutes)
- Presentation: Snowflake data sharing features
- Exercise: Creating your first data share and consuming data from another data share
- Q&A
Snowflake Marketplace (20 minutes)
- Presentation: Introduction to Snowflake Marketplace
- Exercise: Using dataset from Marketplace
- Q&A
- Break (5 minutes)
Using Snowflake internal DB and dashboard creation (25 minutes)
- Presentation: Snowflake metadata walkthrough
- Exercise: Creating your first Snowsight Dashboard
- Q&A
Programmability (20 minutes)
- Presentation: How to run custom code in Snowflake
- Exercise: Create your first Stored procedure or user defined function
- Q&A
Snowpark & Streamlit (25 minutes)
- Presentation: Introduction into Snowpark for Python and Streamlit
- Exercise: Use Snowpark for Python for data transformation
- Q&A
Serverless features (25 minutes)
- Presentation: Snowflake serverless features
- Exercise: Serverless Task creation
- Q&A
- Break (5 minutes)
Performance tuning (30 min)
- Presentation: How to use query profile
- Exercise: Analyzing some long running queries using query plan
- Q&A and Wrap Up
Your Instructor
Tomáš Sobotík
Tomáš Sobotík is a seasoned data architect, specializing in cloud solutions within the AWS ecosystem and the Snowflake platform. He’s a Snowflake Data Superhero and certified Snowflake architect. A technology enthusiast and passionate data developer, he has over 14 years of experience working on BI- and data-related projects spanning various industries.