Scuba requires that the data you wish to import is living in an Azure Storage Container, please follow our docs for complete guidance on setting this up.

  1. How To Create A Storage Account

  2. How To Create A Container

  3. How To Access Storage Account Keys

Before importing data into Scuba Lite, it is also important to understand best practices for data structure and formatting for your data. We highly recommended you read through the documents linked below in an effort to reduce the number of transformations necessary when going through the import process. Please reach out to Talk to an Expert if there are any questions.

  1. Data Set Requirement Checklist

  2. Best Practices for Formatting Data for Ingest

  3. What You Should Know About Structuring Your Data

  4. What to Think About Before You Add Data

Data Pipeline and Import Overview

How to Conduct a Data Import for Scuba Lite

  1. The first step in importing data into Scuba will require you to download and edit a version of our config.json file found HERE. Please review and complete the config import file then resume the following steps below.

  2. In the Azure Portal search for Managed Applications or select the Managed Applications icon from the Azure Services bar.

  3. Select the Application Name of the Scuba Deployment you completed from the Deployment steps.

  4. Under the Essentials tab, click the link for the managed resource group your Scuba Deployment lives under.

  5. Select the resource ending with the word import (ex: scbcvtxvkikhjoc4import) with a type of “Storage Account.”

  6. On the left navigation bar under “Data Storage” click on “Containers.”

  7. Click on the “table-config” container.

  8. Click “Upload” in the top toolbar.

  9. Select the config.json file you created from Step 1 in the import process.

  10. Once uploaded you will see the file populate inside the table-config container.

  11. To confirm and monitor the file import process, go back to the managed resource group from Step 6 and select the resource ending with ADF (ex: scbcvtxvkikhjoc4ADF) with a type of “Data Factory v2.”

  12. Under “Getting Started” click “Open Azure Data Factory Studio.”

  13. Once ADF loads, open the Monitoring utility on the left navigation panel to the third button down.

  14. Here you will see the pipeline name uploaded from the config.json file from Step 1 and determine what status that pipeline is currently in.

  15. While in the Scuba UI, press refresh and confirm that the table populates from the table selection dropdown on the top bar of the Explore view or Data Tab.