Databricks
Create a Databricks fabric to connect Prophecy to your existing Databricks workspace. With a Databricks fabric, you can connect to existing Spark clusters or create new ones, run Spark pipelines, and read or write data, depending on your Databricks permissions.
The following sections describe the parameters needed to set up a Databricks fabric.
Databricks Runtime 16.4 supports both Scala 2.12 and 2.13, but defaults to Scala 2.13. Databricks Runtime 17.0 and later only support Scala 2.13.
If your cluster defaults to Scala 2.13 but your Prophecy installation uses libraries built for Scala 2.12, you might see the following error:
Library installation attempted on the driver node of cluster 0805-161652-bmq6jitu and failed. Cannot resolve Maven library coordinates. Verify library details, repository access, or Maven repository availability. Error code: ERROR_MAVEN_LIBRARY_RESOLUTION, error message: Library resolution failed because unresolved dependency: io.prophecy:prophecy-libs_2.13:3.5.0-8.11.1: not found
To fix this, update Prophecy to version 4.2.0.1 or later, which adds support for Scala 2.13.
Basic Info
The Basic Info tab includes the following parameters:
| Parameter | Description |
|---|---|
| Name | Name used to identify the project. |
| Description (Optional) | Description of the project. |
| Team | Each fabric is associated with one team. All team members will be able to access the fabric in their projects. |
Providers
The Providers tab lets you configure the execution environment settings.
| Parameter | Description |
|---|---|
| Provider Type | Type of fabric to create (in this case, Spark). |
| Provider | Provider of the execution environment (in this case, Databricks). |
Credentials
Fill out the credentials section to verify your Databricks credentials.
| Parameter | Description |
|---|---|
| Databricks Workspace URL | The URL that points to the workspace that the fabric will use as the execution environment. |
| Authentication Method | The method Prophecy will use to authenticate Databricks connections. Access level is tied to the authenticated user’s permissions. At minimum, the authenticated user must have permission to attach clusters in Databricks to use the connection in Prophecy. Some policies additionally require Databricks Workspace Admin permissions. |