Databricks Plugin

Overview

This is a PDI plugin. Deploying this plugin will enable the use of the Bulk load into Databricks Job Entry. This lets users load large volumes of data from files in their cloud accounts into Databricks tables.

Installation

Pre-Installation

  • Ensure that you have access to the Support Portalarrow-up-right. Reach out to your Pentaho Support representative if you do not have access.

  • Ensure that any Pentaho Data Integration (PDI) tools (pan, kitchen, Carte, Spoon) and Pentaho Server are fully shut down before beginning the installation.

Installation Instructions

1

Download the Plugin

2

Extract the files

Unzip the dowloaded file

3

Run the Installer

  • Execute the appropriate script based on your operating system:

    • install.sh for Linux/macOS

    • install.bat for Windows

  • Accept the End User License Agreement (EULA) when prompted

  • Choose Installation Path

    • For PDI/Spoon: <PENTAHO_HOME>/data-integration/plugins

    • For Pentaho Server: <PENTAHO_HOME>/pentaho-server/pentaho-solutions/system/kettle/plugins

    • Note: <PENTAHO_HOME> is the absolute path of the folder where you have installed Pentaho

Post-installation Steps

  • Restart PDI and/or Pentaho Server after installation.

  • Sample transformations are included in the plugin release folder for reference.

Further Details

See Bulk load into Databricks Job

Last updated

Was this helpful?