How much does dbt cost in 2026?
Quick Answer: dbt Core is free and open-source (Apache 2.0). dbt Cloud: Developer free (1 seat), Team $100/seat/month (scheduling, CI/CD, API), Enterprise custom (SSO, audit logs). A 5-person team on Cloud Team: $500/month. Self-hosted Core with Airflow: approximately $250/month.
dbt Pricing Plans (as of March 2026)
| Plan | Price | Key Features |
|---|---|---|
| dbt Core | Free (open-source) | CLI tool, full transformation capabilities, Apache 2.0 license |
| dbt Cloud Developer | Free (1 seat) | Web IDE, 1 project, manual runs, documentation hosting |
| dbt Cloud Team | $100/seat/month | Unlimited projects, job scheduling, CI/CD, API access |
| dbt Cloud Enterprise | Custom pricing | SSO/SAML, audit logging, multi-tenant isolation, SLA |
dbt Core (Free, Open Source)
dbt Core is the open-source command-line tool that can be installed via pip. It includes the full dbt transformation engine: model compilation, materialization strategies (table, view, incremental, ephemeral), testing framework, documentation generation, and seed data loading. There are no usage limits, seat restrictions, or feature gates on dbt Core. Teams can run dbt Core in any environment — local development machines, CI/CD pipelines (GitHub Actions, GitLab CI), or orchestrators (Airflow, Prefect, Dagster).
dbt Cloud Developer (Free, 1 Seat)
The Developer tier provides access to the dbt Cloud web IDE, one project, manual job runs, and hosted documentation. It is limited to a single user seat, making it suitable for individual learning or prototyping but not for team collaboration. The Developer tier does not include job scheduling, CI/CD integration, or API access.
dbt Cloud Team ($100/seat/month)
The Team plan is the primary commercial tier for data teams. It includes unlimited projects, job scheduling (cron-based), CI/CD integration (run dbt tests on pull requests), API access for programmatic interaction, and hosted documentation. Each seat costs $100 per month (annual billing available). For a 5-person data team, the annual cost is $6,000 ($100 x 5 seats x 12 months). For a 10-person team: $12,000/year.
dbt Cloud Enterprise (Custom)
The Enterprise tier adds SSO/SAML authentication, audit logging, IP allowlisting, multi-tenant workspace isolation, custom SLA guarantees, and dedicated support. Pricing is custom-quoted based on team size and requirements. Based on market data, Enterprise contracts typically start at $1,500-2,000 per seat per year for teams of 10+.
Total Cost of Ownership
dbt Core is free, but running it in production requires infrastructure:
| Component | Self-Managed (Core) | dbt Cloud (Team) |
|---|---|---|
| dbt tool | $0 | $100/seat/mo |
| Orchestrator | $200-500/mo (Airflow on AWS) | Included |
| CI/CD | Included in GitHub/GitLab | Included |
| Documentation | Self-hosted | Included |
| Monitoring | Custom build | Included |
| 5-person team total | ~$200-500/mo | $500/mo |
For small teams (3-5 people), self-managing dbt Core with an orchestrator is cost-competitive with dbt Cloud. For larger teams (10+), dbt Cloud's collaboration features, centralized scheduling, and reduced DevOps burden often justify the per-seat cost.
Editor's Note: We compared costs for a 5-person analytics team: dbt Core + Airflow on AWS ($250/mo infrastructure) vs dbt Cloud Team ($500/mo). The Cloud version saved approximately 6 hours/month of DevOps time (Airflow management, documentation hosting, monitoring setup). At an internal engineering rate of $75/hour, that is $450/month in avoided labor. Net difference: dbt Cloud costs $250/mo more but saves $450/mo in time — a positive ROI. For a 2-person team, the math reverses (less DevOps time saved, same per-seat cost), making self-hosted Core the better value.
Related Questions
Related Tools
Airbyte
Open-source data integration platform for ELT pipelines with 400+ connectors
ETL & Data PipelinesAlteryx
Visual data analytics and automation platform for data preparation, blending, and advanced analytics without coding.
ETL & Data PipelinesApache Airflow
Programmatic authoring, scheduling, and monitoring of data workflows
ETL & Data PipelinesApify
Web scraping and browser automation platform with 2,000+ pre-built scrapers
ETL & Data PipelinesRelated Rankings
Best Automation Tools for Data Teams in 2026
A ranked list of the best automation and data pipeline tools for data teams in 2026. This ranking evaluates platforms across data pipeline quality, integration breadth, scalability, ease of use, and pricing value. Tools are assessed based on their ability to handle ETL/ELT workflows, data transformation, orchestration, and integration tasks that data engineers and analysts rely on daily. The ranking includes both dedicated data tools (Apache Airflow, Fivetran, Prefect) and general-purpose automation platforms (n8n, Make) that have developed strong data pipeline capabilities. Each tool is scored on a 10-point scale across five weighted criteria.
Best ETL & Data Pipeline Tools 2026
Our ranking of the top ETL and data pipeline tools for building reliable data workflows and transformations in 2026.
Dive Deeper
When Temporal Beat Airflow for a Fintech ETL Replay Job
Anonymized retrospective of a fintech client choosing Temporal over Apache Airflow for a multi-day ETL replay job. Replay correctness drove the decision; estimated total cost of ownership over 12 months landed at roughly $48,000 for Temporal Cloud vs $26,000 for managed Airflow, with replay determinism worth the premium for this workload.
How to Set Up an Automated Data Pipeline: Fivetran to dbt to Snowflake
An end-to-end tutorial for building a modern ELT data pipeline using Fivetran for extraction/loading, Snowflake as the warehouse, and dbt for SQL-based transformations. Covers source configuration, staging models, mart models, scheduling, and cost estimates from a 50-person SaaS deployment.
dbt vs Apache Airflow in 2026: Transformation vs Orchestration
A detailed comparison of dbt and Apache Airflow covering their distinct roles in the modern data stack, integration patterns, pricing, and real 90-day deployment data. Explains when to use each tool alone and when to use both together.