Snowflake

Store your event and dispatch data in Snowflake through S3 integration

Snowflake Data Warehouse Integration

The Snowflake Data Warehouse integration provides a powerful solution for storing and analyzing your event and dispatch data. We utilize S3 as an intermediary storage layer, allowing you to easily import the data into Snowflake's cloud data platform.

How the Integration Works

  • Daily S3 Deposits: Events and dispatches are automatically deposited into your S3 bucket on a daily basis
  • Complete Data: All events and dispatches from your account are included in the deposits
  • Efficient Loading: Snowflake's COPY command efficiently loads data from S3
  • Scalable Analytics: Snowflake's architecture enables fast querying of large datasets

Data Organization

The data in S3 is organized in a partitioned structure:

Please reference our S3 documentation for that.

Setting Up Snowflake Access to S3

Snowflake provides several methods to access S3 data. We recommend reviewing their official documentation for detailed setup instructions.

Data Loading

Once configured, you can load data into Snowflake using:

COPY INTO your_table
FROM @your_stage
FILE_FORMAT = (TYPE = 'PARQUET')
PATTERN = '.*events/.*.parquet';

Key Features

Snowflake's integration enables powerful data processing capabilities:

  • Automatic Scaling: Resources scale automatically based on workload
  • Zero-Copy Cloning: Create instant copies of your data
  • Time Travel: Access historical data versions
  • Secure Sharing: Share data with other Snowflake accounts
  • Materialized Views: Optimize query performance

Best Practices

  • Use Snowflake's recommended file formats and compression
  • Implement appropriate access controls
  • Monitor your S3 and Snowflake usage
  • Take advantage of the partitioning structure for efficient loading
  • Consider using Snowpipe for continuous data loading