The Lake Loader is an application that loads Snowplow events to a cloud storage bucket using Open Table Formats.
The Lake Loader on AWS is a fully streaming application that continually pulls events from Kinesis and writes to S3.
The Lake Loader is published as a Docker image which you can run on any AWS VM. You do not need a Spark cluster to run this loader.
docker pull snowplow/lake-loader-aws:0.2.0
To run the loader, mount your config files into the docker image, and then provide the file paths on the command line:
docker run \
For some output formats, you need to pull a slightly different tag to get a compatible docker image. The configuration reference page explains when this is needed.
Configuring the loader
The loader config file is in HOCON format, and it allows configuring many different properties of how the loader runs.
The simplest possible config file just needs a description of your pipeline inputs and outputs:
See the configuration reference for all possible configuration parameters.
"Windowing" is an important config setting, which controls how often the Lake Loader commits a batch of events to the data lake. If you adjust this config setting, you should be aware that data lake queries are most efficient when the size of the parquet files in the lake are relatively large.
- If you set this to a low value, the loader will write events to the lake more frequently, reducing latency. However, the output parquet files will be smaller, which will make querying the data less efficient.
- Conversely, if you set this to a high value, the loader will generate bigger output parquet files, which are efficient for queries — at the cost of events arriving to the lake with more delay.
The default setting is
5 minutes. For moderate to high volumes, this value strikes a nice balance between the need for large output parquet files and the need for reasonably low latency data.
"windowing": "5 minutes"
If you tune this setting correctly, then your lake can support efficient analytic queries without the need to run an
OPTIMIZE job on the files.
The Lake Loader requires an Iglu resolver file which describes the Iglu repositories that host your schemas. This should be the same Iglu configuration file that you used in the Enrichment process.
By default, Snowplow collects telemetry data for Lake Loader (since version 0.1.0). Telemetry allows us to understand how our applications are used and helps us build a better product for our users (including you!).
This data is anonymous and minimal, and since our code is open source, you can inspect what’s collected.
If you wish to help us further, you can optionally provide your email (or just a UUID) in the
telemetry.userProvidedId configuration setting.
If you wish to disable telemetry, you can do so by setting
See our telemetry principles for more information.