Supported warehouse and data lake destinations
Data warehouses and data lakes are primary destinations for Snowplow data. For other options, see the destinations overview page.
How loading works
The Snowplow data loading process is engineered for large volumes of data. In addition, for each data warehouse, our loader applications ensure the best representation of Snowplow events. That includes automatically adjusting the database types for self-describing events and entities according to their schemas.
For more details on the loading flow, pick a destination below and follow the link in the Loader column, where you will find additional information and diagrams.
Data warehouse loaders
The cloud selection is for where your Snowplow pipeline runs. The warehouse itself can be deployed in any cloud.
- AWS
- GCP
- Azure
Destination | Type | Loader application | Status |
---|---|---|---|
Redshift (including Redshift serverless) | Batching (recommended) or micro-batching | RDB Loader | Production-ready |
BigQuery | Streaming | BigQuery Loader | Production-ready |
Snowflake | Streaming | Snowflake Streaming Loader | Production-ready |
Databricks | Batching (recommended) or micro-batching | Snowplow RDB Loader | Production-ready |
Databricks | Streaming | Databricks Streaming Loader | Early release |
Destination | Type | Loader application | Status |
---|---|---|---|
BigQuery | Streaming | BigQuery Loader | Production-ready |
Snowflake | Streaming | Snowflake Streaming Loader | Production-ready |
Databricks | Micro-batching (via a data lake) | Lake Loader | Production-ready |
Databricks | Streaming | Databricks Streaming Loader | Early release |
Destination | Type | Loader application | Status |
---|---|---|---|
BigQuery | Streaming | BigQuery Loader | Production-ready |
Snowflake | Streaming | Snowflake Streaming Loader | Production-ready |
Databricks | Micro-batching (via a data lake) | Lake Loader | Production-ready |
Databricks | Streaming | Databricks Streaming Loader | Early release |
Synapse Analytics | Micro-batching (via a data lake) | Lake Loader | Production-ready |
Data lake loaders
All lake loaders are micro-batching.
- AWS
- GCP
- Azure
Lake | Format | Compatibility | Loader application | Status |
---|---|---|---|---|
S3 | Delta | Athena, | Lake Loader | Production-ready |
S3 | Iceberg | Athena, Redshift | Lake Loader | Production-ready |
S3 | TSV/JSON | Athena | S3 Loader | Only recommended for use with RDB Batch Transformer or for raw failed events |
Please note that currently the S3 Delta loader is not compatible with Databricks. The loader uses DynamoDB tables for mutually exclusive writes to S3, a feature of Delta. Databricks, however, does not support this (as of July 2025). This means that it’s not possible to alter the data via Databricks (e.g. to run OPTIMIZE
or to delete PII).
Lake | Format | Compatibility | Loader application | Status |
---|---|---|---|---|
GCS | Delta | Databricks | Lake Loader | Production-ready |
GCS | JSON | BigQuery | GCS Loader | Only recommended for raw failed events |
Lake | Format | Compatibility | Loader application | Status |
---|---|---|---|---|
ADLS Gen2 | Delta | Synapse Analytics, Fabric, Databricks | Lake Loader | Production-ready |