Welcome to dbtvault!¶
dbtvault is a dbt package that generates & executes the ETL you need to build a Data Vault 2.0 Data Warehouse.
You need to be running dbt to use the package. If needed, you can find get more guidance on how to get set up from the dbt documentation.
Go check them out!
dbt is designed for ease of use in data engineering: for when you need to develop a data pipeline.
dbt offers a command-line utility developed in Python that can run on your desktop or inside a VM in your network and is free to download and use. Alternatively, you can use their SaaS offering dbt Cloud which functions as a dbt IDE.
Our package runs inside the dbt environment, so you can use dbt to run other parts of the Data Vault pipeline, combined with the dbtvault package for the Data Vault 2.0 specific steps.
Join our Slack community!
What is Data Vault 2.0?¶
Data Vault 2.0 is an Agile method that can be used to deliver a highly scalable enterprise Data Warehouse.
The method covers the full approach for developing a Data Warehouse: architecture, data modelling, development, and includes a number of unique techniques.
If you want to learn about Data Vault 2.0, your best starting point is the book Building a Scalable Data Warehouse with Data Vault 2.0 (see details below).
Why do Data Vault 2.0 and dbt integrate well?¶
The Data Vault 2.0 method uses a small set of standard building blocks to model your data warehouse (Hubs, Links and Satellites in the Raw Data Vault) and, because they are standardised, you can load these blocks with templated SQL. dbt allows for a template-driven implementation using Jinja. This leads to better quality code, fewer mistakes, and greatly improved productivity: i.e. Agility.
Where does dbtvault fit in?¶
The dbtvault package generates and runs Data Vault ETL code from your metadata (table names and mapping details) which is then provided to your dbt models contains calls to dbtvault macros. The macro does the rest of the work: it processes the metadata, generates SQL and then dbt executes the load respecting any and all dependencies.
dbt even runs the load in parallel. As Data Vault 2.0 is designed for parallel load and Snowflake is highly parallelised, your ETL load will finish in rapid time. Your experience may vary form platform to platform, however we aim to be as consistent as possible.
dbtvault reduces the need to write SQL by hand to load the Data Vault, which is a repetitive, time-consuming and potentially error-prone task.
What are the advantages of dbtvault?¶
dbt works with the dbtvault package to:
- Generate SQL to process the staging layer and load the data vault.
- Ensure consistency and correctness in the generated SQL.
- Identify dependencies between SQL statements.
- Create Raw Data Vault tables when a release first identifies them.
- Execute all generated SQL statements as a complete set.
- Execute data load in parallel up to a user-defined number of parallel threads.
- Generate data flow diagrams showing data lineage.
- Automatically build a documentation website.
If you are going to use the dbtvault package for your Data Vault 2.0 project, then we expect you to have some prior knowledge about the Data Vault 2.0 method.
In the case you do need to get up to speed with Data Vault 2.0, we recommend the following resources for further information:
Books (from Amazon)¶
- Building a Scalable Data Warehouse with Data Vault 2.0, Dan Linstedt and Michael Olschimke
- The Data Vault Guru: a pragmatic guide on building a data vault, Patrick Cuba
- Better Data Modelling: An Introduction to Agile Data Engineering Using Data Vault 2.0, Kent Graziano
Blogs and Downloads¶
Roadmap and Changelog¶
We keep an up-to-date log of past and planned changes: