Skip to main content

Dimensional Data Modeling

1h 37m 57s
English
Paid

Course description

In today's world, where data plays a key role, effective organization of information is the foundation for quality analytics and report building. Multidimensional data modeling is an important approach that allows structuring data for quick access and informed decision-making.

This course provides a detailed introduction to the fundamental concepts of dimensional modeling. You will learn how fact and dimension tables work, what slowly changing dimensions (SCD) are, as well as the different types of fact tables.

During the course, you will gain practical experience in setting up and working with a data warehouse using real tools - DuckDB and DBeaver. By the end of the training, you will confidently understand how to design a data model for high-performance analytics and reporting.

Read more about the course

Course includes:

Introduction to Data Warehouses

You will learn the basics of building warehouses and their importance for analytical processing. Discover how a data warehouse consolidates information from various sources for fast and scalable analysis.

Basics of Dimensional Modeling

You will become acquainted with key elements: dimensional and fact tables. You will learn how to design a data structure for analytics, determine business metrics, find suitable dimensions, and link them together. Through practical examples, you will master the principles of building a meaningful and efficient data model.

Setting Up a Data Warehouse

You will learn to work with DuckDB and DBeaver: create tables, manage them, and prepare the environment for analytics. The module includes step-by-step instructions for setting up the environment and deploying your first warehouse.

Working with a Data Warehouse

Dive into advanced topics: learn to handle slowly changing dimensions (SCD), work with different types of fact tables - transactional and cumulative. You will understand how to track business events and trends, as well as optimize queries and analyze large volumes of data.

Watch Online

This is a demo lesson (10:00 remaining)

You can watch up to 10 minutes for free. Subscribe to unlock all 16 lessons in this course and access 10,000+ hours of premium content across all courses.

View Pricing
0:00
/
#1: Introduction

All Course Lessons (16)

#Lesson TitleDurationAccess
1
Introduction Demo
01:58
2
Course Goals
02:06
3
Intro to Data Warehousing
06:43
4
Approaches to building a data warehouse
05:21
5
Dimension tables explained
05:35
6
Fact tables explained
06:35
7
Identifying dimensions
03:17
8
What is duckdb
05:59
9
First DuckDB hands-on
02:21
10
Creating tables in duckdb
02:41
11
Installing dbeaver
06:50
12
Exploring scd0 and scd1
19:58
13
Exploring scd2
13:53
14
Exploring transaction fact table
06:29
15
Exploring accumulating fact table
07:18
16
Conclusion
00:53

Unlock unlimited learning

Get instant access to all 15 lessons in this course, plus thousands of other premium courses. One subscription, unlimited knowledge.

Learn more about subscription

Comments

0 comments

Want to join the conversation?

Sign in to comment

Similar courses

Data Engineering on GCP

Data Engineering on GCP

Sources: Andreas Kretz
Google Cloud Platform (GCP) is one of the most popular cloud platforms in the world, providing an extensive set of tools and services for building...
1 hour 17 minutes 33 seconds
Snowflake for Data Engineers

Snowflake for Data Engineers

Sources: Andreas Kretz
Snowflake is a next-generation cloud data warehouse that everyone is talking about today. The platform operates 100% in the cloud, providing flexible access...
2 hours 4 minutes 8 seconds
Modern Data Warehouses & Data Lakes

Modern Data Warehouses & Data Lakes

Sources: Andreas Kretz
As a data engineer, you will regularly work with analytics platforms where companies store data in Data Lakes and Data Warehouses for building...
58 minutes 9 seconds
The Data Engineering Bootcamp: Zero to Mastery

The Data Engineering Bootcamp: Zero to Mastery

Sources: zerotomastery.io
Learn to build streaming pipelines with Apache Kafka and Flink, create data lakes on AWS, run ML workflows on Spark, and integrate LLM models into...
16 hours 46 minutes 22 seconds
Data Analysis for Beginners: Python & Statistics

Data Analysis for Beginners: Python & Statistics

Sources: zerotomastery.io
This course is your first step into the world of data analysis using one of the main tools for analysts - Python. Without complicated terms, advanced...
6 hours 34 minutes 20 seconds