Configure the AEL daemon for a Hive service
You must configure the application.properties file of the AEL daemon if you want to:
Use Hive tables on a secure supported HDP cluster.
Use Hive managed and unmanaged tables in an ORC or Parquet format on your Amazon EMR cluster.
Use Hive managed and unmanaged tables in an ORC or Parquet format on your Google Dataproc cluster.
To configure the properties file, perform the following steps.
Navigate to the
data-integration/adaptive-execution/configdirectory and open theapplication.propertiesfile with any text editor.Set the values for your environment as shown in the following table.
ParameterValueenableHiveConnectionEnables AEL access to Hive tables. Set this value to
true.spark.driver.extraClassPathSpecifies the path to the directory containing the
hive-site.xmlfile on the driver node. It loads thehive-site.xmlfile as a resource in the driver. This resource defines the Hive endpoints and security setting required by AEL to access the Hive subsystem.spark.executor.extraClassPathSpecifies the path to the directory containing the
hive-site.xmlon the executor nodes. It loads thehive-site.xmlfile as a resource on each executor. This resource defines the Hive endpoints and security setting required by AEL to access the Hive subsystem.The following lines of code show sample values for these parameters:
# AEL Spark Hive Property Settings enableHiveConnection=true enableHiveWarehouseConnector=false spark.driver.extraClassPath=/etc/spark/conf.dist/ spark.executor.extraClassPath=/etc/spark/conf.dist/Save and close the file.
Restart the AEL daemon.
You can now use PDI with a Hadoop Hive 2/3 database connection.
Last updated
Was this helpful?

