Snowflake
Loading data into Snowflake via external stages.
This guide will walk you through the process of exporting data from your DCN into Snowflake using Google Cloud Storage (GCS), Amazon S3, or Azure Blob Storage as an intermediary storage bucket. This leverages Snowflake's external stage feature, which allows you to load data stored in an external stage into a Snowflake table, which can then be re-loaded into your DCN via the Snowflake source.
Prerequisites
Before proceeding, ensure that you have the following:
An active Snowflake account with necessary permissions.
An active Google Cloud Storage (GCS), Amazon S3, or Azure Blob Storage bucket where data from your DCN will be exported to.
Step 1 - Export Data from your DCN
The first step is to export the data from your DCN into the previously created Google Cloud Storage (GCS), Amazon S3, or Azure Blob Storage bucket in Parquet format. Once your audience is exported, you can then proceed to step 2.

Step 2 - Configure your external stage & load your data.
Amazon S3:
Allow Access: Configure permissions and allow access to S3.
Configure S3: Follow Snowflake's guide for configuring an S3 integration.
Create S3 Stage: Create an S3 stage in Snowflake.
Copy Data: Execute the COPY command for S3.
Google Cloud Storage (GCS):
Configure GCS: Follow Snowflake's guide for configuring a GCS integration.
Copy Data: Execute the COPY command for GCS.
Azure Blob Storage:
Allow Access: Configure permissions and allow access to Azure.
Configure Azure: Follow Snowflake's guide for configuring an Azure integration.
Create Azure Stage: Create an Azure stage in Snowflake.
Copy Data: Execute the COPY command for Azure.
Step 3 - Query Data in Snowflake
With the data successfully copied into Snowflake's internal tables (as detailed in Step 2), you are now ready to execute queries on these tables just like any other Snowflake table. To enhance your experience, consider utilizing Snowflake's features such as indexing, clustering, or other performance optimization techniques, especially when dealing with large datasets.
Last updated
Was this helpful?