This page provides you with instructions on how to extract data from Microsoft SQL Server and load it into Google BigQuery. (If this manual process sounds onerous, check out Stitch, which can do all the heavy lifting for you in just a few clicks.)
What is Microsoft SQL Server?
Microsoft SQL Server is a relational database management system that supports applications on a single machine, on a local area network, or across the web. SQL Server supports Microsoft's .NET framework out of the box, and integrates nicely into the Microsoft ecosystem.
What is Google BigQuery?
Google BigQuery is a data warehouse that delivers super-fast results from SQL queries, which it accomplishes using a powerful engine dubbed Dremel. With BigQuery, there's no spinning up (and down) clusters of machines as you work with your data. With that said, it's clear why some claim that BigQuery prioritizes querying over administration. It's super fast, and that's the reason why most folks use it.
Getting data out of SQL Server
The most common way most folks who work with databases get their data is by using queries for extraction. With SELECT statements you can filter, sort, and limit the data you want to retrieve. If you need to export data in bulk, you can use Microsoft SQL Server Management Studio, which enables you to export entire tables and databases in formats like text, CSV, or SQL queries that can restore the database if run.
Loading data into Google BigQuery
Google Cloud Platform offers a helpful guide for loading data into BigQuery. You can use the
bq command-line tool, and in particular the
bq load command, to upload files to your datasets, adding schema and data type information along the way. You can find the syntax in the Quickstart guide for bq. Iterate through this process as many times as it takes to load all of your tables into BigQuery.
Keeping SQL Server data up to date
All set! You've written a script to move data from SQL Server into your data warehouse. But data freshness is one of the most important aspects of any analysis – what happens when you have new data that you need to add?
You could load the entire SQL Server database again. Doing this is almost guaranteed to be slow and painful, and cause all kinds of latency.
A better approach is to build your script to recognize new and updated records in the source database. Using an auto-incrementing field as a key is a great way to accomplish this. The key functions something like a bookmark, so your script can resume where it left off. When you've built in this functionality, you can set up your script as a cron job or continuous loop to get new data as it appears in SQL Server.
Other data warehouse options
BigQuery is great, but sometimes you need to optimize for different things when you're choosing a data warehouse. Some folks choose to go with Amazon Redshift, PostgreSQL, Snowflake, or Microsoft Azure Synapse Analytics, which are RDBMSes that use similar SQL syntax, or Panoply, which works with Redshift instances. Others choose a data lake, like Amazon S3 or Delta Lake on Databricks. If you're interested in seeing the relevant steps for loading data into one of these platforms, check out To Redshift, To Postgres, To Snowflake, To Panoply, To Azure Synapse Analytics, To S3, and To Delta Lake.
Easier and faster alternatives
If all this sounds a bit overwhelming, don’t be alarmed. If you have all the skills necessary to go through this process, chances are building and maintaining a script like this isn’t a very high-leverage use of your time.
Thankfully, products like Stitch were built to move data from Microsoft SQL Server to Google BigQuery automatically. With just a few clicks, Stitch starts extracting your Microsoft SQL Server data, structuring it in a way that's optimized for analysis, and inserting that data into your Google BigQuery data warehouse.