Connecting to Databricks

Prev Next

Overview

Azure Databricks provides an open analytics platform designed for creating, deploying, sharing, and maintaining large-scale data, analytics, and AI solutions. The Databricks Data Intelligence Platform connects with your cloud account’s storage and security, handling cloud infrastructure management and deployment. Syniti Knowledge Platform (SKP) allows you to connect to Databricks on Azure for the following usage types:

  • Data Quality

  • Replication

  • Profiling

Connection Properties

When connecting to a Databricks database from the Syniti Knowledge Platform (SKP), provide the following connection properties to establish a secure and reliable connection. This section explains each property to understand what information is required and how it is used.

Property

Description

Notes

Databricks Host
(Required)

The network endpoint (URL or hostname) of the Databricks workspace where the SQL endpoint or cluster is hosted.

This identifies the server to which clients send connection requests.

HTTP Path
(Required)

The specific REST/SQL API path that uniquely identifies the SQL warehouse (previously “SQL endpoint”) or compute context within Databricks.

This path, combined with the host, determines the exact service your client will communicate with.

Port
(Required)

The TCP port number used for client connections.

For Databricks ODBC connections this is typically 443 (standard HTTPS) unless otherwise specified.

Use SSL

A boolean setting that determines whether the connection uses Secure Sockets Layer (SSL)/TLS encryption.

When on, communication is encrypted to protect data in transit between the client and Databricks.

Thrift Transport
(Required)

Specifies the transport protocol used by the driver for communication. Common options include Binary or HTTP transport modes.

This influences how SQL commands are sent and results received at the protocol level.

Database

The default database context to use after connecting.

This determines where unqualified object names (e.g., table references without schema) are resolved.

Schema

The logical grouping within the selected database that organizes tables, views, and other objects.

Setting the schema identifies the namespace for object resolution after connecting.

Catalog

The catalog defines a higher-level namespace that can contain multiple databases and schemas.

In Databricks Unity Catalog environments, it determines which catalog scope is used for object access and privileges.

Authentication Configuration

Authentication Mechanism

Description

Notes

0

No Authentication

Not recommend; use only if your cluster allows it.

2

Username Authentication

Legacy; UID only, no password in many configuration.

3

Personal Access Token

Recommended for most API to Server scenarios.

11

OAuth 2.0

Microsoft Entra ID Authentication, with different AuthFlow values.

Authentication Flow

Used with OAuth 2.0 and Microsoft Entra ID - AuthMech=11

Flow Value

Description

0

Token Pass-through. Using secret token.

1

Machine-to-Machine. Using clientID and secret.

3

Azure managed identity authentication

Related Articles

Create a Connection