Configure the AEL daemon for a Hive service
You must configure the application.properties
file of the AEL daemon if you want to:
Use Hive tables on a secure supported HDP cluster.
Use Hive managed and unmanaged tables in an ORC or Parquet format on your Amazon EMR cluster.
Use Hive managed and unmanaged tables in an ORC or Parquet format on your Google Dataproc cluster.
To configure the properties file, perform the following steps.
Navigate to the
data-integration/adaptive-execution/config
directory and open theapplication.properties
file with any text editor.Set the values for your environment as shown in the following table.
ParameterValueenableHiveConnection
Enables AEL access to Hive tables. Set this value to
true
.spark.driver.extraClassPath
Specifies the path to the directory containing the
hive-site.xml
file on the driver node. It loads thehive-site.xml
file as a resource in the driver. This resource defines the Hive endpoints and security setting required by AEL to access the Hive subsystem.spark.executor.extraClassPath
Specifies the path to the directory containing the
hive-site.xml
on the executor nodes. It loads thehive-site.xml
file as a resource on each executor. This resource defines the Hive endpoints and security setting required by AEL to access the Hive subsystem.The following lines of code show sample values for these parameters:
# AEL Spark Hive Property Settings enableHiveConnection=true enableHiveWarehouseConnector=false spark.driver.extraClassPath=/etc/spark/conf.dist/ spark.executor.extraClassPath=/etc/spark/conf.dist/
Save and close the file.
Restart the AEL daemon.
You can now use PDI with a Hadoop Hive 2/3 database connection.
Last updated
Was this helpful?