Azure Storage and Snowflake DW Integration

Snowflake is one of the most popular cloud data warehouses today . Snowflake has a number of features that make it an amazing cloud data warehouse to work with for data engineers. Snowflake leverages several new Azure features, including limitless storage accounts, accelerated networking, and storage soft delete. The goal of Snowflake is to provide the same Snowflake experience no matter which cloud infrastructure provider customers choose, with no barriers to entry. Snowflake on Azure will make it easier than ever for teams across an organization to become more data-driven, efficient and productive with their most valuable resources: data and people.

In this article I show you how to setup a notification integration in Snowflake that talks to Azure event grid linked to an Azure storage account. Essentially this service will listen to any events (file additions in this case) in an Azure Storage container and perform a given activity via Snowpipe. Here is a high level architecture for this setup.

Step 1

Setup an Azure storage account by following the Microsoft docs here. Once a a storage account is setup, create a container as shown below.

Next, we need to create a Queue as shown below. Then note down the queue URL.

Next, lets setup an event subscription by first clicking on the Events icon in the left pane. Below are the configuration details for the event subscription in azure.

With that you should now have the following setup at the Azure end.

  1. Azure storage account
  2. a container named ‘myfolder’
  3. a Queue called ‘snowpipe-notification-queue’
  4. a notification event configured to use the queue from item 3

Step 2

Log into your snowflake account (if you don’t have one you can create a 30 day free trial account here) and open a brand new worksheet and now its time to do some coding!

Step 3

Execute the following code in your Snowflake worksheet to create a notification integration. You would need your Azure tenant id and the queue url for this piece of code to work. A notification integration is a Snowflake object that provides an interface between Snowflake and third-party cloud message queuing services.

create notification integration SNOWPIPE_DEMO_EVENT
  enabled = true
  type = queue
  notification_provider = azure_storage_queue
  azure_storage_queue_primary_uri = 'https://yourazurestorageaccountname.queue.core.windows.net/snowpipe-notification-queue'
  azure_tenant_id = 'your azure tenant id';

Next, run the code below to find the value for AZURE_CONSENT_URL

 desc integration SNOWPIPE_DEMO_EVENT

Step 4

Open the URL in a new tab, you will be prompted to login to your azure portal and allow the app registration to be added to your azure tenant. Then navigate back to you Azure storage account and assign the “Storage Queue Data Contributor” role to the app registration as shown below.

Step 5

Next, we create a stage object in snowflake

  CREATE OR REPLACE STAGE "DEMO_DB"."PUBLIC"."DEMO_STAGE"
  url = 'azure://yourazurestorageaccountname.blob.core.windows.net/myfolder/'
  credentials = (azure_sas_token=
    'your sas token from your azure storage account'
  );

Step 6

For the above code you would need a SAS token. You can navigate to the “Shared access signature” page in your Azure Storage account and generate a key for the container as per your security requirements. refer here to learn more.

Run the code below to make sure you can see a list of file in the container or just an empty result if you don’t have files yet. But make sure you don’t get any errors when you run this.

 ls @"DEMO_DB"."PUBLIC"."DEMO_STAGE";

Step 7

Now it time to finally create the snowpipe object


  CREATE OR REPLACE pipe "DEMO_DB"."PUBLIC"."DEMO_PIPE"
  auto_ingest = true
  integration = 'SNOWPIPE_DEMO_EVENT'
  as
  copy into employees
  from @"DEMO_DB"."PUBLIC"."DEMO_STAGE"
  file_format = my_csv_format--(type = 'CSV')

Above code will be executed each time a new file is added to the azure storage container. For this code to work make sure you have a table created. In my case I have an ’employee’ table created which has a structure shown below.

 create table employees
  (
  firstname varchar(100),
  lastname varchar(100),
  empid varchar(100)
  );

the raw csv file I am using to load into this table is below

Now you should see the whole process in action as soon as you drop a csv file into the container in Azure. Please feel free to email me if you have questions

Published by Narayan Sujay Somasekhar

• 12+ years of experience leading the build of BI and Cloud Data Platform solutions using cloud technologies such as Snowflake, Azure Synapse, Databricks and AWS Redshift. • Over 8+ years as a Data Analytics and Engineering practice leader with demonstrated history of working with management consulting firms across Tax & Accounting, Finance, Power & Utility industry. • Experience in managing the team roadmap, and delivering actionable data insights to sales, product, marketing, and senior leadership. • Strong background in Data Technology Solutions delivery & Data Automation for business processes using various tools. • Expertise in bringing Data-Driven IT Strategic Planning to align metrics, communicate data changes across reporting, Enterprise Data Warehouses, Data Lakes and Customer Relationship Managements Systems. • Experienced working with cross functional teams, Data Scientists/Analysts and Business Managers in building Data Science and Data Engineering practice from the ground up. • Experienced in Designing and implementing NLP solutions with focus on sentiment analysis, opinion mining, key phase extraction using Azure Cognitive Services and Amazon Comprehend • Extensive programming experience with SQL, Python, C#, R, and Scala.

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s

%d bloggers like this: