Snowflake
Store your event and dispatch data in Snowflake through S3 integration
Snowflake Data Warehouse Integration
The Snowflake Data Warehouse integration provides a powerful solution for storing and analyzing your event and dispatch data. We utilize S3 as an intermediary storage layer, allowing you to easily import the data into Snowflake's cloud data platform.
How the Integration Works
- Daily S3 Deposits: Events and dispatches are automatically deposited into your S3 bucket on a daily basis
- Complete Data: All events and dispatches from your account are included in the deposits
- Efficient Loading: Snowflake's COPY command efficiently loads data from S3
- Scalable Analytics: Snowflake's architecture enables fast querying of large datasets
Data Organization
The data in S3 is organized in a partitioned structure:
Please reference our S3 documentation for that.
Setting Up Snowflake Access to S3
Snowflake provides several methods to access S3 data. We recommend reviewing their official documentation for detailed setup instructions.
Data Loading
Once configured, you can load data into Snowflake using:
COPY INTO your_table
FROM @your_stage
FILE_FORMAT = (TYPE = 'PARQUET')
PATTERN = '.*events/.*.parquet';
Key Features
Snowflake's integration enables powerful data processing capabilities:
- Automatic Scaling: Resources scale automatically based on workload
- Zero-Copy Cloning: Create instant copies of your data
- Time Travel: Access historical data versions
- Secure Sharing: Share data with other Snowflake accounts
- Materialized Views: Optimize query performance
Best Practices
- Use Snowflake's recommended file formats and compression
- Implement appropriate access controls
- Monitor your S3 and Snowflake usage
- Take advantage of the partitioning structure for efficient loading
- Consider using Snowpipe for continuous data loading
Updated about 19 hours ago