Databricks
- UpdatedJul 31, 2025
- 1 minute read
- Zurich
- Data and Automation
The Databricks connector provides read-only access to data and metadata from the Databricks data analytics platform.
A connection admin can set up a connection to Databricks in the Workflow Data Fabric Hub and grant data stewards access to this connection. Data stewards can then use the established connection to create a data fabric table and map data from Databricks. This allows users to access Databricks data through the table list view or by using GlideRecord scripts. For details on creating data fabric tables and mapping data, see Managing data fabric tables in Workflow Data Fabric Hub.
The connector has been enhanced to improve the performance of the following Glide queries and list view operations. These improvements allow the majority of queries to be executed at the data source.
- Sort
- Limit
- Filter
- GroupBy
- avg()
- count()
- max()
- min()
- sum()
- References
Supported data types
The following table lists supported Databricks data types and the default matching data types in a data fabric table.
| Databricks | Data fabric table |
|---|---|
| bigint | Long Integer String |
| int | Long Integer String |
| smallint | Long Integer String |
| tinyint | Long Integer String |
| timestamp_ntz | Basic Date/Time |
| char | Char |
| timestamp | Basic Date/Time |
| date | Date |
| decimal | Decimal |
| string | String |
| boolean | True/False |
| double | Floating Point Number |
| float | Floating Point Number |
Related Content
- Create a Databricks connection
Establish a zero copy connection to the Databricks data analytics platform in Workflow Data Fabric Hub.