Welcome to dbtvault!¶
dbtvault is a dbt package that generates & executes the ETL you need to run a Data Vault 2.0 Data Warehouse on a Snowflake database.
You need to be running dbt to use the package.
Go check them out!
dbt is designed for ease of use in data engineering: for when you need to develop a data pipeline. It is a single command-line utility that can run on your desktop or a VM in your network, it is developed in Python, and it is free to download and use.
Our package runs inside the dbt environment, so you can use dbt to run other parts of the Data Vault pipeline, combined with the dbtvault package for the Data Vault specific steps.
Sign up for early-bird announcements or join our Slack¶
What is Data Vault 2.0?¶
Data Vault 2.0 is an Agile method that can be used to deliver a highly scalable enterprise Data Warehouse.
The method covers the full approach for developing a Data Warehouse: architecture, data modelling, development, and includes a number of unique techniques.
If you want to learn about Data Vault 2.0, your best starting point is the book Building a Scalable Data Warehouse with Data Vault 2.0 (see details below).
Data Vault 2.0 supports code automation.¶
Essentially, the method uses a small set of standard building blocks to model your data warehouse ( Hubs, Links and Satellites in the Raw Data Vault) and, because they are standardised, you can load these blocks with templated SQL. The result is a template-driven implementation, populated by metadata. You provide the metadata (table names and mapping details) and SQL is generated automatically. This leads to better quality code, fewer mistakes, and greatly improved productivity: i.e. Agility.
What does dbtvault do?¶
The dbtvault package generates and runs Data Vault ETL code from your metadata.
Just like other dbt projects, you write a model for each ETL step. You provide the metadata for each model as declared variables and include code to invoke a macro from the dbtvault package. The macro does the rest of the work: it processes the metadata, generates Snowflake SQL and then dbt executes the load respecting any and all dependencies.
dbt even runs the load in parallel. As Data Vault 2.0 is designed for parallel load and Snowflake is highly performant, your ETL load will finish in rapid time.
dbtvault reduces the need to write Snowflake SQL by hand to load the Data Vault, which is is a repetitive, time-consuming and potentially error prone task.
What features does dbt running the dbtvault package offer?¶
dbt works with the dbtvault package to:
- Generate SQL to process the staging layer and load the data vault.
- Ensures consistency and correctness in the generated SQL.
- Identify dependencies between SQL statements.
- Create Raw Data Vault tables when a release first identifies them.
- Execute all generated SQL statements as a complete set.
- Execute data load in parallel up to a user-defined number of parallel threads.
- Generate data flow diagrams showing data lineage.
- Automatically build a documentation website.
You Do Need Some Prior Knowledge About the Data Vault 2.0 Method¶
If you are going to use the dbtvault package for your Data Vault 2.0 project, then we expect you to have some prior knowledge about the Data Vault 2.0 method.
How Can I Get Up to Speed on Data Vault 2.0?¶
You can get further information about the Data Vault 2.0 method from the following resources:
Books (from Amazon)¶
- Building a Scalable Data Warehouse with Data Vault 2.0, Dan Linstedt and Michael Olschimke
- The Data Vault Guru: a pragmatic guide on building a data vault, Patrick Cuba
- Better Data Modelling: An Introduction to Agile Data Engineering Using Data Vault 2.0, Kent Graziano
Blogs and Downloads¶
Roadmap and Changelog¶
We keep an up-to-date log of past and planned changes: