Course by:

Course Highlights

  • Submitted to NCVET for NSQF Alignment.
  • The Data Engineering Workshop (DNGW) is the fifth workshop in the Snowflake’s Hands-on Essentials Workshop series.The workshop gives learners extended practice querying data prior to load and the iterative development of File Formats using that querying process. Learners also learn about GeoSpatial data and use many of Snowflake’s built-in geospatial functions.
  • The workshop requires hands-on lab work to earn a badge. The lab work is auto-graded.
  • Skill Type

  • Course Duration

  • Domain

  • GOI Incentive applicable

  • Course Category

  • Nasscom Assessment

  • Placement Assistance

  • Certificate Earned

  • Content Alignment Type

  • NOS Details

  • Mode of Delivery

Course Details

Learning Objectives

What will you learn in Snowflake Hands-on Essentials Series Part 5 - Data Engineering?

The Snowflake Hands-on Essentials Series - Data Engineering focuses on providing practical skills for building and managing data pipelines within the Snowflake environment. Here's a breakdown of the key learning objectives:

  • Learn how to efficiently load data into Snowflake from various sources, including cloud storage (S3, Azure Blob Storage, Google Cloud Storage).
  • Master the use of the COPY INTO command for bulk data loading.
  • Understand and implement Snowpipe for continuous data ingestion.
  • Gain proficiency in using SQL for data transformation and manipulation within Snowflake.
  • Learn how to create and utilize views, stored procedures, and user-defined functions (UDFs) for data processing.
  • Understand and implement data transformation using Snowflake Streams and Tasks for change data capture (CDC) and automation.
  • Learn how to automate data pipelines using Snowflake Tasks.
  • Understand how to schedule and manage data processing workflows.
  • Learn how to use Snowpark for pipeline creation.
  • Understand the importance of data quality in data engineering.
  • Learn techniques for data validation and cleansing within Snowflake.
  • Learn how to use data profiling techniques.
  • Learn how to optimize data pipelines for performance and efficiency.
  • Understand how to use clustering and partitioning to improve query performance.
  • Learn how to monitor and troubleshoot data pipeline performance.
  • Gain proficiency in handling and querying semi-structured data formats (JSON, XML) within Snowflake.
  • Understand how to use variant columns.
  • Learn to use SnowSQL CLI.
  • Understand the use of external tables.
  • Develop hands-on experience through practical labs and exercises.
  • Build functional data pipelines that demonstrate acquired skills.
Read more
Reasons to enrol

Why should you take Snowflake Hands-on Essentials Series Part 5 - Data Engineering course?

At the end of this course, the learner will be able use Snowflake to:

  • Work with timezone functions and time data types.
  • Create and run tasks.
  • Create and run pipes.
  • Create, edit and run dashboards.
  • Use a SQL MERGE statement.
  • Transform data during a COPY INTO.
  • Use Cloud provider topics and queues to trigger loading processes.
Read more
Ideal Participants

Who should take Snowflake Hands-on Essentials Series Part 5 - Data Engineering course?

  • Designed for people new to Snowflake or new to database work in general. The course can be used by managers who simply want to understand what Snowflake is generally capable of, or it can be used by those considering a career or as a data professional. Likewise, seasoned data professionals find the courses in this series a quick and easy introduction to tasks they are already familiar with, in a tool they are not.
Read more
Curriculum

Curriculum for the Snowflake Hands-on Essentials Series Part 5 - Data Engineering Course

  • Data Ingestion:
    • Loading data from various sources.
    • Working with internal and external stages.
  • Data Transformation:
    • SQL-based transformations.
    • Snowpark for data transformations using programming languages.
    • User-defined functions (UDFs).
    • Snowflake Streams for change data capture (CDC).
    • Stored procedures for complex logic.
    • Dynamic tables.
  • Data Pipeline Orchestration:
    • Automating workflows with Snowflake Tasks.
    • Building directed acyclic graphs (DAGs) for task dependencies.
  • Data Delivery:
    • Data sharing through the Snowflake Marketplace.
    • Utilizing Streamlit in Snowflake Applications.
    • Snowflake Native applications.
  • Key Skills and Concepts:
    • Time zones and timestamp formats.
    • CTAS statements.
    • Parsing JSON with paths and casts.
    • Session settings and query profiles.
    • SQL MERGE statements.
    • SQL window functions.
    • Metadata storage.
  • Practical Learning:
    • Hands-on labs and exercises.
    • Automated lab grading with DORA.
    • Real-world scenarios and use cases.
  • Read more
    skills and tools

    Tools you will learn in Snowflake Hands-on Essentials Series Part 5 - Data Engineering course

    • Set role, namespace and warehouse USER defaults.
    • Create database, schema, file format.
    • Create and test a Stage.
    • Run a COPY INTO.
    • Parse JSON with paths and casts.
    • Create a View.
    • Use current_timestamp function.
    • Edit a view.
    • Set session timezone.
    • View account parameters (timezone).
    • Use the ILIKE operator.
    • Use the wildcard operator.
    • Use the PARSE_IP function.
    • Use a function with a JSON path.
    • Locate a data share and add it to your account.
    • Add usage privilege on a share to a lower role.
    • Make use of functions provided from a share.
    • Join local data with shared data in a select.
    • View Query Profiles.
    • Use functions in a join.
    • Use the CONVERT_TIMEZONE function.
    • Use the DAYNAME function.
    • Create a table with SQL.
    • Insert rows in a table with SQL.
    • Use the LISTAGG function.
    • Use an unspecified date or time function to join to a lookup table and return a value from it.
    • Create a table and load it via a CTAS command.
    • Create a Task.
    • Grant task execution to SYSADMIN role.
    • Execute a Task Manually.
    • View task history via the GUI.
    • Create Merge statements.
    • Truncate tables.
    • Create a Resource Monitor.
    • Suspend and Resume Tasks.
    • View Task History.
    • Trace the flow of records through a pipeline.
    • Use task dependencies (root and non-root) in task definitions.
    • View task dependencies graph.
    • Use serverless compute option for a task.
    • Add file metadata columns to rows when loading data.
    • Transform data during the COPY INTO.
    • Create a Snowpipe.
    • Use auto-ingest in a Snowpipe.
    • Use SNS Topic property in a Snowpipe.
    • Create and view a simple stream.
    • Use a merge to process stream data.
    • Add a Stream dependency to task schedule.
    • Create a dashboard.
    • Add a tile to a dashboard.
    • Create a bar chart for a dashboard.
    • Create a heat map for a dashboard.
    • Use a WINDOW function to roll up records.
    Read more