Advanced analytics engineering with Snowflake and dbt

In this one-day workshop, you will learn how to create and orchestrate modern data transformation pipelines with dbt and Snowflake. You will discover how you can improve your pipelines’ performance and quality as well as optimize costs by choosing right materialization techniques, version controlling, data testing, monitoring, and scheduling. We will walk you through typical data transformation problems you can encounter on a journey to deliver fresh & reliable data and how modern tooling can help to solve them. All hands-on exercises will be carried out in a public cloud environment (e.g. GCP or AWS).

   Target Audience

Data analysts, analytics engineers & data engineers, who are interested in learning how to build and deploy production-ready Snowflake data pipelines using state-of-the-art data management solutions.

    Requirements

  • SQL and Python fluency: ability to write data transforming queries and scripts
  • Basic understanding of ETL processes
  • Basic experience with a command-line interface
  • Laptop with a stable internet connection (participants will connect to pre-created cloud development environment)

    Participant’s ROI

  • Concise and practical knowledge of applying dbt to solve typical problems with data pipelines in Snowflake: managing run sequence, data quality issues, monitoring, and orchestrating transformations.
  • Hands-on coding experience under the supervision of Data Engineers experienced in maintaining dbt pipelines on Snowflake production environments
  • Tips about real-world applications and best practices.

    Training Materials

During the workshop, participants will follow a shared step-by-step guideline and perform a series of hands-on exercises on their Snowflake free trial accounts. Cloud IDE environments will be supplied for each participant. Appropriately prepared datasets will be provided to use for all participants to take part in the example real-life use case scenario.

    Time Box

1 Day event

    Agenda

Session #1 - Introduction to data pipelines with Snowflake & dbt

  • Snowflake ecosystem - key functionalities & new features
  • Core concepts of dbt (recap)
    Data models
  • Seeds, sources
  • Tests
  • Documentation
  • Transforming data using SQL with dbt
  • Hands-on exercises

Session #2 - Simple, production-ready end-do-end data pipeline

  • Data ingestion with modern data integration tools
  • Data discovery (data search, usage statistics, data lineage)
  • Data profiling & exploration
  • Working with different materialization techniques in Snowflake (including Snowflake Dynamic Tables)
  • Workflow scheduling & orchestration
  • Hands-on exercises

Session #3 - Data quality testing & monitoring

  • Data testing using dbt tests
  • Data monitoring with Snowflake Alerts & Notifications
  • (Bonus task) Building sample visualizations of Snowflake data products
  • Hands-on exercises

    Session leader:

Data Architect / Technical Product Owner
GetInData | Part of Xebia
Data Platform Architect
GetInData | Part of Xebia

BIG DATA TECHNOLOGY
WARSAW SUMMIT 2024

10-11th of April 2024

ORGANIZER

Evention sp. z o.o

Rondo ONZ 1 Str,

Warsaw, Poland

www.evention.pl

CONTACT

Weronika Warpas

© 2024 | This site uses cookies. By continuing to browse the site you are agreeing to our use of cookies.