Installation
Table of Contents
Pre-requisites
- Install
Databricks CLI
- Ensure that you have the Databricks Command-Line Interface (CLI) installed on your machine. Refer to the installation instructions provided for Linux, MacOS, and Windows, available here.
Installing the Databricks CLI in different OS:
- MacOS & Linux
- Windows
- Linux without brew


#!/usr/bin/env bash
#install dependencies
apt update && apt install -y curl sudo unzip
#install databricks cli
curl -fsSL https://raw.githubusercontent.com/databricks/setup-cli/v0.242.0/install.sh | sudo sh
- Configure
Databricks CLI
- Details can be found here. Additionally, Lakebridge requires the profile used for the Databricks CLI to specify a cluster_id, to do this, you can either:
- Edit your
~/.databrickscfg
file directly and enter acluster_id
for the profile you're using or - The flag
--configure-cluster
gives you the prompt to select the cluster_id from the available clusters on the workspace specified on the selected profile.
databricks configure --host <host> --configure-cluster --profile <profile_name>
- Alternatively you can use the environment variable
DATABRICKS_CLUSTER_ID
to set the cluster id you would want to use for your profile before running thedatabricks configure
command.
export DATABRICKS_CLUSTER_ID=<cluster_id>
databricks configure --host <host> --profile <profile_name>
Python
- Verify that Python 3.10 or above is installed
Windows
- Install python from here. Your Windows computer will need a shell environment (GitBash or WSL)MacOS/Unix
- Use brew to install python in macOS/Unix machines
Check Python version on Windows, macOS, and Unix

Java
- Verify that Java 11 or above is installed. This is required for the Morpheus transpiler
Install Lakebridge
Upon completing the environment setup, install Lakebridge by executing the following command:
databricks labs install lakebridge
This will install Lakebridge using the workspace details set in the DEFAULT profile. If you want to install it using a different profile, you can specify the profile name using the --profile
flag.
databricks labs install lakebridge --profile <profile_name>
To view all the profiles available, you can run the following command:
databricks auth profiles

Verify Installation
Verify the successful installation by executing the provided command; confirmation of a successful installation is indicated when the displayed output aligns with the example below:
databricks labs lakebridge --help

Install Transpile
Upon completing the environment setup, you can install the out of the box transpilers by executing the following command. This command will also prompt for the required configuration elements so that you don't need to include them in your command-line call every time.
databricks labs lakebridge install-transpile

Override the default[Bladebridge] config:
There is an option for you to override the default config file that Lakebridge
uses for converting source code from dialects like informatica
, datastage
, synapse
,
orcale
etc. During installtion you may use your own custom config file and Lakebridge
will override the config with the one you would provide. You can only setup this
override during installation.
Specify the config file to override the default[Bladebridge] config during installation:
Specify the config file to override the default[Bladebridge] config - press <enter> for none (default: <none>): <local_full_path>/custom_<source>2databricks.json
Verify Installation
Verify the successful installation by executing the provided command; confirmation of a successful installation is indicated when the displayed output aligns with the example screenshot provided:
databricks labs lakebridge transpile --help

Configure Reconcile
Once you're ready to reconcile your data, you need to configure the reconcile module.
databricks labs lakebridge configure-reconcile

SQL Warehouse for Reconcile
While configuring the reconcile properties, lakebridge by default creates a SQL warehouse. lakebridge uses user profile to authenticate to any Databricks resource and hence
if the user running this command doesn't have permission to create SQL warehouse, the configure-reconcile would fail. In this case users can provide the
warehouse_id
of an already created SQL warehouse that they have atleast CAN_USE permission on in the databricks profile (~/.databrickscfg
) using which they
are running the lakebridge commands and lakebridge would use that warehouse to complete the reconcile configuration instead of trying to create a new one.
This is how the profile would look like:
[profile-name]
host = <your-workspace-url>
...
warehouse_id = <your-warehouse-id>
Verify Configuration
Verify the successful configuration by executing the provided command; confirmation of a successful configuration is indicated when the displayed output aligns with the example screenshot provided:
databricks labs lakebridge reconcile --help
