Set the Spark parameters globally
Within the application.properties file, you may add any number of Spark properties to make global changes to the application tuning parameters for your Spark cluster that runs PDI. To view the full list of Spark parameters, see Spark properties documentation. .
Spark tuning may be affected by the following factors:
When a Hadoop or a Spark cluster is a shared enterprise asset.
When cluster resources are shared among many Spark applications that are processed in parallel.
Perform the following steps to set up the application.properties file:
Log on to the cluster and stop the AEL daemon as described in Step 6 of Configure the AEL daemon for YARN mode.
Navigate to the
design-tools/data-integration/adaptive-execution/configfolder and open theapplication.propertiesfile with any text editor.Enter the Spark configuration parameter and value for each setting that you want to make in the cluster. For example,
spark.yarn.executor.memoryOverhead=1024Note: See Determining Spark resource requirements for an example of calculating resources.
Save and close the file.
Restart the AEL daemon as described in Step 6 of Configure the AEL daemon for YARN mode.
The Spark parameters configured in the properties file are now globally applied to the Spark cluster. The performance results of your executed transformations are available on the YARN ResourceManager and Spark History Server. You can refine the tuning of the cluster at the transformation level as described in Set the Spark parameters locally in PDI.
Last updated

