Page History
This is a guide to installing the WhereScape Enablement Pack for Databricks for WhereScape RED.
Table of Contents maxLevel 2 minLevel 2
Include Page | ||||
---|---|---|---|---|
|
Prerequisites For Target Database Databricks
- Before you begin the following prerequisites must be met:
- Create Database and ODBC DSN:
- Databricks (ODBC driver version 2.7.5 or higher(64-bit))
- At least one schema available to use as a RED Data Warehouse Target
- Databricks (ODBC driver version 2.7.5 or higher(64-bit))
- Software Installations
- Databricks CLI - Refer to Setup Guide Databricks CLI Setup
- Python 3.8 or higher
- Select Add Python 3.8 to PATH from the installation Window
- Pip Manager Install with the command:
python -m pip install --upgrade pip
Include Page | ||||
---|---|---|---|---|
|
Post Install Steps - Optional
If you used the script Setup_Enablement_Pack.ps1
then the following optional post-install steps are available
Configure Connections
There were Three connections added that will optionally require your attention:
- Connection: 'Database Source System' - this connection was set up as an example source connection,
- open it's properties and set it up for a source DB in your environment
- or you can remove it if not required.
- Connection: 'Databricks' - this connection was set as per parameters provided in script 1
- open properties and check ifthe Database ID is set correctly
- open its properties and check the extended properties tab, set it up for HTTP_PATH, SERVER_HOSTNAME, DB_ACCESS_TOKEN, and DBFS_TMP
Enable Script Launcher Toolbar
Several stand-alone scripts provide some features such as "Ranged Loading", these scripts have been added to the Script Launcher menu but you will need to enable the menu toolbar item to see them.
To enable the Script Launcher menu in RED: Select menu item View>Toolbars>Script Launcher
Source Enablement Pack Support
Source Pack Name | Supported by Databricks | Supported Features | Prerequisites |
Amazon S3 | Yes | Bulk load to Databricks | Include the Access Key and Secret Key in the Amazon S3 Cloud Parser Connection for S3. For guidance on obtaining these credentials, please refer to the relevant documentation: {+}https://docs.aws.amazon.com/IAM/latest/UserGuide/security-creds.html+ |
Azure Data Lake Storage Gen2 | Yes | Bulk load to Databricks | Add the SAS Token to the ADLG2 Cloud Parser Connection. Refer to {+}https://learn.microsoft.com/en-us/azure/storage/common/storage-sas-overview+ for information on SAS Tokens. |
Google Cloud Storage | Yes | Bulk load to Databricks | Step 1: Service Account Setup
|
Windows Parser | 1. CSV | Load Template, Source Properties will have the option to select parser type to load the files. | Refer to Windows Parser Guide |
Include Page | ||||
---|---|---|---|---|
|