How to set up a data pipeline with Fivetran
Quick Answer: Fivetran automates data pipeline creation by connecting to source systems, replicating data to a destination warehouse, and maintaining schema consistency with zero code. Add a connector, authenticate the source, select a destination, choose the sync frequency, and start the initial sync.
How to Set Up a Data Pipeline with Fivetran
Fivetran is a managed ELT (Extract, Load, Transform) platform that replicates data from sources to cloud data warehouses without custom code. As of April 2026, Fivetran supports 500+ connectors and integrates with Snowflake, BigQuery, Redshift, Databricks, and other destinations. Pricing is usage-based on Monthly Active Rows (MAR), starting at $1/credit on the Standard plan.
Step 1: Add a Connector
Log into the Fivetran dashboard and click "Add Connector." Search for the data source from the connector library:
- SaaS applications: Salesforce, HubSpot, Stripe, Shopify, Google Analytics, Jira
- Databases: PostgreSQL, MySQL, SQL Server, MongoDB, Oracle
- Files: S3, Google Cloud Storage, Azure Blob, SFTP
- Events: Segment, Webhooks, Kafka
Step 2: Authenticate the Source
Each connector has a specific authentication method:
- OAuth (Salesforce, HubSpot, Google): Click "Authorize" and sign in
- API key (Stripe, Intercom): Paste the API key from the source application
- Database credentials (PostgreSQL, MySQL): Provide host, port, database name, username, and password. Fivetran requires read-only access and recommends a dedicated service account
Fivetran provides setup guides for each connector with exact permissions and configuration steps.
Step 3: Select a Destination
Choose the data warehouse where Fivetran will load the data. Configure the destination once (Fivetran remembers it for all connectors):
- Snowflake: Provide account URL, database, warehouse, and role
- BigQuery: Provide project ID and dataset (Fivetran creates tables automatically)
- Redshift: Provide cluster endpoint, database, and credentials
- Databricks: Provide workspace URL, token, and catalog
Step 4: Choose Schema and Tables
Fivetran displays the source schema and allows selection of which tables to sync. Options:
- Sync all tables — Replicate the entire source schema
- Select specific tables — Choose only the tables needed
- Column hashing — Hash sensitive columns (PII) during replication
Fivetran maintains schema consistency: if the source adds a column, Fivetran automatically adds it to the destination.
Step 5: Set Sync Frequency
Configure how often Fivetran checks for new and changed data:
- 5 minutes — Near real-time (highest MAR usage)
- 15 minutes — Balanced freshness and cost
- 1 hour — Standard for reporting workloads
- 6 hours / 24 hours — Batch analytics with lower cost
Step 6: Start the Initial Sync
Click "Save & Test" to validate the connection, then start the initial sync. The first sync replicates all historical data and may take hours or days depending on data volume. Subsequent syncs are incremental, transferring only new and changed records.
Post-Setup: Connect to dbt for Transformations
Fivetran loads raw data into the warehouse. Use dbt (data build tool) to transform raw tables into analytics-ready models. Fivetran offers a native dbt Cloud integration that triggers dbt runs automatically after each sync completes.
Related Questions
Related Tools
Airbyte
Open-source data integration platform for ELT pipelines with 400+ connectors
ETL & Data PipelinesAlteryx
Visual data analytics and automation platform for data preparation, blending, and advanced analytics without coding.
ETL & Data PipelinesApache Airflow
Programmatic authoring, scheduling, and monitoring of data workflows
ETL & Data PipelinesApify
Web scraping and browser automation platform with 2,000+ pre-built scrapers
ETL & Data PipelinesRelated Rankings
Best Automation Tools for Data Teams in 2026
A ranked list of the best automation and data pipeline tools for data teams in 2026. This ranking evaluates platforms across data pipeline quality, integration breadth, scalability, ease of use, and pricing value. Tools are assessed based on their ability to handle ETL/ELT workflows, data transformation, orchestration, and integration tasks that data engineers and analysts rely on daily. The ranking includes both dedicated data tools (Apache Airflow, Fivetran, Prefect) and general-purpose automation platforms (n8n, Make) that have developed strong data pipeline capabilities. Each tool is scored on a 10-point scale across five weighted criteria.
Best ETL & Data Pipeline Tools 2026
Our ranking of the top ETL and data pipeline tools for building reliable data workflows and transformations in 2026.
Dive Deeper
How to Set Up an Automated Data Pipeline: Fivetran to dbt to Snowflake
An end-to-end tutorial for building a modern ELT data pipeline using Fivetran for extraction/loading, Snowflake as the warehouse, and dbt for SQL-based transformations. Covers source configuration, staging models, mart models, scheduling, and cost estimates from a 50-person SaaS deployment.
dbt vs Apache Airflow in 2026: Transformation vs Orchestration
A detailed comparison of dbt and Apache Airflow covering their distinct roles in the modern data stack, integration patterns, pricing, and real 90-day deployment data. Explains when to use each tool alone and when to use both together.
Airbyte vs Fivetran in 2026: Open-Source vs Managed ELT
A data-driven comparison of Airbyte and Fivetran covering architecture, connector ecosystems, pricing at scale, reliability, compliance certifications, and real 60-day parallel deployment results. Covers self-hosted, cloud, and enterprise options for both platforms.