Using Spark Submit

You can use the Spark Submit job entry, along with an externally developed Spark execution script, to run Spark jobs on your YARN clusters.

The following instructions explain how to use the Spark Submit entry in PDI to run a Spark job. To learn how to execute Spark Submit jobs on secure Cloudera CDP clusters, see Use Kerberos with Spark Submit in the Administer Pentaho Data Integration and Analytics document.

Note: Before you start, you must install and configure the Spark client according to the instructions for the Spark Submit job entry in the Pentaho Data Integration document.

Last updated

Was this helpful?