Skip to main content

Best practices for workflows

This page contains the collective wisdom of experienced users of dbt on how to best use it in your analytics work. Observing these best practices will help your analytics team work as effectively as possible, while implementing the pro-tips will add some polish to your dbt projects!

Best practice workflows

Version control your dbt project

All dbt projects should be managed in version control. Git branches should be created to manage development of new features and bug fixes. All code changes should be reviewed by a colleague (or yourself) in a Pull Request prior to merging into master.

Git guide

We've codified our best practices in Git, in our Git guide.

Use separate development and production environments

dbt makes it easy to maintain separate production and development environments through the use of targets within a profile. We recommend using a dev target when running dbt from your command line and only running against a prod target when running from a production deployment. You can read more about managing environments here.

Use a style guide for your project

SQL styles, field naming conventions, and other rules for your dbt project should be codified, especially on projects where multiple dbt users are writing code.

Our style guide

We've made our style guide public – these can act as a good starting point for your own style guide.

Best practices in dbt projects

Use the ref function

The ref function is what makes dbt so powerful! Using the ref function allows dbt to infer dependencies, ensuring that models are built in the correct order. It also ensures that your current model selects from upstream tables and views in the same environment that you're working in. Always use the ref function when selecting from another model, rather than using the direct relation reference (e.g. my_schema.my_table).

Limit references to raw data

Your dbt project will depend on raw data stored in your database. Since this data is normally loaded by third parties, the structure of it can change over time – tables and columns may be added, removed, or renamed. When this happens, it is easier to update models if raw data is only referenced in one place.

Using sources for raw data references

We recommend defining your raw data as sources, and selecting from the source rather than using the direct relation reference. Our dbt projects don't contain any direct relation references in any models.

Rename and recast fields once

Raw data is generally stored in a source-conformed structure, that is, following the schema and naming conventions that the source defines. Not only will this structure differ between different sources, it is also likely to differ from the naming conventions you wish to use for analytics.

The first layer of transformations in a dbt project should:

  • Select from only one source
  • Rename fields and tables to fit the conventions you wish to use within your project, for example, ensuring all timestamps are named <event>_at. These conventions should be declared in your project coding conventions (see above).
  • Recast fields into the correct data type, for example, changing dates into UTC and prices into dollar amounts.

All subsequent data models should be built on top of these models, reducing the amount of duplicated code.

What happened to base models?

Earlier versions of this documentation recommended implementing “base models” as the first layer of transformation, and gave advice on the SQL within these models. We realized that while the reasons behind this convention were valid, the specific advice around "base models" represented an opinion, so we moved it out of the official documentation.

You can instead find our opinions on how we structure our dbt projects.

Break complex models up into smaller pieces

Complex models often include multiple Common Table Expressions (CTEs). In dbt, you can instead separate these CTEs into separate models that build on top of each other. It is often a good idea to break up complex models when:

  • A CTE is duplicated across two models. Breaking the CTE into a separate model allows you to reference the model from any number of downstream models, reducing duplicated code.
  • A CTE changes the grain of a the data it selects from. It's often useful to test any transformations that change the grain (as in, what one record represents) of your data. Breaking a CTE into a separate model allows you to test this transformation independently of a larger model.
  • The SQL in a query contains many lines. Breaking CTEs into separate models can reduce the cognitive load when another dbt user (or your future self) is looking at the code.

Group your models in directories

Within your models/ directory, you can have any number of nested subdirectories. We leverage directories heavily, since using a nested structure within directories makes it easier to:

  • Configure groups of models, by specifying configurations in your dbt_project.yml file.
  • Run subsections of your DAG, by using the model selection syntax.
  • Communicate modeling steps to collaborators
  • Create conventions around the allowed upstream dependencies of a model, for example, "models in the marts directory can only select from other models in the marts directory, or from models in the staging directory".

Add tests to your models

dbt provides a framework to test assumptions about the results generated by a model. Adding tests to a project helps provide assurance that both:

  • your SQL is transforming data in the way you expect, and
  • your source data contains the values you expect
Recommended tests

Our style guide recommends that at a minimum, every model should have a primary key that is tested to ensure it is unique, and not null.

Consider the information architecture of your data warehouse

When a user connects to a data warehouse via a SQL client, they often rely on the names of schemas, relations, and columns, to understand the data they are presented with. To improve the information architecture of a data warehouse, we:

  • Use custom schemas to separate relations into logical groupings, or hide intermediate models in a separate schema. Generally, these custom schemas align with the directories we use to group our models, and are configured from the dbt_project.yml file.
  • Use prefixes in table names (for example, stg_, fct_ and dim_) to indicate which relations should be queried by end users.

Choose your materializations wisely

materialization determine the way models are built through configuration. As a general rule:

  • Views are faster to build, but slower to query compared to tables.
  • Incremental models provide the same query performance as tables, are faster to build compared to the table materialization, however they introduce complexity into a project.

We often:

  • Use views by default
  • Use ephemeral models for lightweight transformations that shouldn't be exposed to end-users
  • Use tables for models that are queried by BI tools
  • Use tables for models that have multiple descendants
  • Use incremental models when the build time for table models exceeds an acceptable threshold

Pro-tips for workflows

Use the model selection syntax when running locally

When developing, it often makes sense to only run the model you are actively working on and any downstream models. You can choose which models to run by using the model selection syntax.

Run only modified models to test changes ("slim CI")

To merge code changes with confidence, you want to know that those changes will not cause breakages elsewhere in your project. For that reason, we recommend running models and tests in a sandboxed environment, separated from your production data, as an automatic check in your git workflow. (If you use GitHub and dbt Cloud, read about how to set up CI jobs.

At the same time, it costs time (and money) to run and test all the models in your project. This inefficiency feels especially painful if your PR only proposes changes to a handful of models.

By comparing to artifacts from a previous production run, dbt can determine which models are modified and build them on top of of their unmodified parents.

dbt run -s state:modified+ --defer --state path/to/prod/artifacts
dbt test -s state:modified+ --defer --state path/to/prod/artifacts

By comparing to artifacts from a previous production run, dbt can determine model and test result statuses.

  • result:fail
  • result:error
  • result:warn
  • result:success
  • result:skipped
  • result:pass

For smarter reruns, use the result:<status> selector instead of manually overriding dbt commands with the models in scope.

dbt run --select state:modified+ result:error+ --defer --state path/to/prod/artifacts
  • Rerun all my erroneous models AND run changes I made concurrently that may relate to the erroneous models for downstream use
dbt build --select state:modified+ result:error+ --defer --state path/to/prod/artifacts
  • Rerun and retest all my erroneous models AND run changes I made concurrently that may relate to the erroneous models for downstream use
dbt build --select state:modified+ result:error+ result:fail+ --defer --state path/to/prod/artifacts
  • Rerun all my erroneous models AND all my failed tests
  • Rerun all my erroneous models AND run changes I made concurrently that may relate to the erroneous models for downstream use
  • There's a failed test that's unrelated to modified or error nodes(think: source test that needs to refresh a data load in order to pass)
dbt test --select result:fail --exclude <example test> --defer --state path/to/prod/artifacts
  • Rerun all my failed tests and exclude tests that I know will still fail
  • This can apply to updates in source data during the "EL" process that need to be rerun after they are refreshed

Note: If you're using the --state target/ flag, result:error and result:fail flags can only be selected concurrently(in the same command) if using the dbt build command. dbt test will overwrite the run_results.json from dbt run in a previous command invocation.

Only supported by v1.1 or newer.

By comparing to a sources.json artifact from a previous production run to a current sources.json artifact, dbt can determine which sources are fresher and run downstream models based on them.

# job 1
dbt source freshness # must be run to get previous state

Test all my sources that are fresher than the previous run, and run and test all models downstream of them:

# job 2
dbt source freshness # must be run again to compare current to previous state
dbt build --select source_status:fresher+ --state path/to/prod/artifacts

To learn more, read the docs on state.

Pro-tips for dbt Projects

Limit the data processed when in development

In a development environment, faster run times allow you to iterate your code more quickly. We frequently speed up our runs by using a pattern that limits data based on the target name:

select
*
from event_tracking.events
{% if target.name == 'dev' %}
where created_at >= dateadd('day', -3, current_date)
{% endif %}

Use hooks to manage privileges on objects that dbt creates

Use grant statements from hooks to ensure that permissions are applied to the objects created by dbt. By codifying these grant statements in hooks, you can version control and repeatably apply these permissions.

Recommended grant statements

We've shared the exact grant statements we use over on Discourse

Separate source-centric and business-centric transformations

When modeling data, we frequently find there are two stages:

  1. Source-centric transformations to transform data from different sources into a consistent structure, for example, re-aliasing and recasting columns, or unioning, joining or deduplicating source data to ensure your model has the correct grain; and
  2. Business-centric transformations that transform data into models that represent entities and processes relevant to your business, or implement business definitions in SQL.

We find it most useful to separate these two types of transformations into different models, to make the distinction between source-centric and business-centric logic clear.

Managing whitespace generated by Jinja

If you're using macros or other pieces of Jinja in your models, your compiled SQL (found in the target/compiled directory) may contain unwanted whitespace. Check out the Jinja documentation to learn how to control generated whitespace.

0