Skip to main content

Supported warehouse and data lake destinations

Data warehouses and data lakes are primary destinations for Snowplow data. For other options, see the destinations overview page.

How loading works

The Snowplow data loading process is engineered for large volumes of data. In addition, for each data warehouse, our loader applications ensure the best representation of Snowplow events. That includes automatically adjusting the database types for self-describing events and entities according to their schemas.

tip

For more details on the loading flow, pick a destination below and follow the link in the Loader column, where you will find additional information and diagrams.

Data warehouse loaders

Cloud

The cloud selection is for where your Snowplow pipeline runs. The warehouse itself can be deployed in any cloud.

DestinationTypeLoader applicationStatus
Redshift
(including Redshift serverless)
Batching (recommended)
or micro-batching
RDB LoaderProduction-ready
BigQueryStreamingBigQuery LoaderProduction-ready
SnowflakeStreamingSnowflake Streaming LoaderProduction-ready
DatabricksBatching (recommended)
or micro-batching
Snowplow RDB LoaderProduction-ready
DatabricksStreamingDatabricks Streaming LoaderEarly release

Data lake loaders

All lake loaders are micro-batching.

LakeFormatCompatibilityLoader applicationStatus
S3DeltaAthena, DatabricksLake LoaderProduction-ready
S3IcebergAthena, RedshiftLake LoaderProduction-ready
S3TSV/JSONAthenaS3 LoaderOnly recommended for use with RDB Batch Transformer or for raw failed events
tip

Please note that currently the S3 Delta loader is not compatible with Databricks. The loader uses DynamoDB tables for mutually exclusive writes to S3, a feature of Delta. Databricks, however, does not support this (as of July 2025). This means that it’s not possible to alter the data via Databricks (e.g. to run OPTIMIZE or to delete PII).

On this page

Want to see a custom demo?

Our technical experts are here to help.