Accessing data from Databricks
You can access data from Databricks in Spotfire.
About this task
Before you begin
- The Databricks connector requires a driver on the computer running Spotfire. See Drivers and data sources in Spotfire.
- To make sure that your database is supported, see the system requirements for the Databricks connector.
Procedure
-
Open the
Files and data
flyout, and click
Connect to.
- In the list of data sources, select Databricks.
- In the panel on the right, choose if you want to create a new connection or add data from a shared data connection:
- Connector for Databricks — Features and settings
You can connect to and access data from Databricks with the data connector for Databricks. On this page, you can find information about the capabilities, available settings, and things to keep in mind when you work with data connections to Databricks.
Working with and troubleshooting Databricks data connections
About this task
Databricks cluster that is not running
When connecting to a Databricks cluster or warehouse that is not already running, the first connection attempt will trigger the resource to start. This can take several minutes. You may have to click Connect again if the connection times out.
Unity Catalog
You can access data from Databricks workspaces that are enabled for
Unity Catalog, which means that data is organized in three levels
(catalog.schema.table). You can browse the data after
connecting, when you select data to include in the connection.
Converting an existing Databricks connection from the Apache Spark SQL connector
The data connector for Databricks was introduced in Spotfire version 14.6.1. In earlier versions, to access data from Databricks, you would use the connector for Apache Spark SQL. If you have existing data connections to Databricks created with the Apache Spark SQL connector, and you want to use the stand-alone Databricks connector instead, replace the data connections in your analyses, see Replacing a data source.