LogoLogo
CtrlK
Try Pentaho Data Integration and Analytics
10.2 Install
  • Pentaho Documentation
  • Install Pentaho Data Integration and Analytics
  • Pentaho installation
    • Acquire and install enterprise licenses
      • Install and manage a local license server
        • Install a local license server on Linux
        • Install a local license server on Windows
        • Administrative commands for local license server
        • Set up Self-Signed Certificate (SSL)
        • Activate offline entitlements for a local license server
        • Uninstall a local license server (Linux)
        • Uninstall a local license server (Windows)
        • Change the temporary directory for Java in Linux
      • Install licenses using PUC
      • Install licenses using the PDI client
      • Manage licenses using the command line interface
    • Archive installation
      • Archive installation process
        • Prepare your Windows environment for an archive install
          • Process overview
          • Create Windows directory structure
          • Install Java
          • Install the Pentaho Repository host database
          • Download and unpack installation files
            • Step 1: Download files
            • Step 2: Unpack the Pentaho Server installation files
            • Step 3: (Optional) Unpack the plugin files
            • Step 4: Verify directory structure
          • Set environment variables
          • Prepare your Pentaho Repository
        • Prepare your Linux environment for an archive install
          • Process overview
          • Create the Pentaho user
          • Create Linux directory structure
          • Install Java
          • Install the Pentaho Repository host database
          • Download and unpack installation files
          • Set environment variables
          • Advanced Linux considerations
            • Systems without video cards
            • Systems without X11
          • Prepare your Pentaho Repository
        • Use PostgreSQL as Your Repository Database (Archive installation)
          • Before you begin
          • Initialize PostgreSQL Pentaho Repository database
            • Step 1: Change default passwords
            • Step 2: Run SQL scripts
            • Step 3: Verify PostgreSQL initialization
          • Configure PostgreSQL Pentaho RepositoryDatabase
            • Step 1: Set up Quartz on PostgreSQL
            • Step 2: Set Hibernate settings for PostgreSQL
            • Step 3: Modify Jackrabbit repository information for PostgreSQL
          • Perform Tomcat-specific connection tasks
            • Step 1: Download driver and apply to the Pentaho Server
            • Step 2: Modify JDBC connection information in the Tomcat XML file
          • Start Your Server
        • Use MySQL or MariaDB as your repository database (Archive installation)
          • Before you begin
          • Initialize MySQL or MariaDB Pentaho Repository database
            • Step 1: Change default passwords
            • Step 2: Run SQL scripts
            • Step 3: Verify MySQL or MariaDB initialization
          • Configure MySQL or MariaDB Pentaho Repository Database
            • Step 1: Set up Quartz on MySQL or MariaDB
            • Step 2: Set Hibernate settings for MySQL or MariaDB
            • Step 3: Replace default version of audit log file with MySQL version
            • Step 4: Modify Jackrabbit repository information for MySQL or MariaDB
          • Perform Tomcat-specific connection tasks
          • Start Your Server
        • Use Oracle as Your Repository Database (Archive installation)
          • Before you begin
          • Initialize Oracle Pentaho Repository database
            • Step 1: Change default passwords
            • Step 2: Run SQL scripts
            • Step 3: Verify Oracle initialization
          • Configure Oracle Pentaho Repository database
            • Step 1: Set up Quartz on Oracle
            • Step 2: Set Hibernate settings for Oracle
            • Step 3: Replace default version of audit log file with Oracle version
            • Step 4: Modify Jackrabbit repository information for Oracle
          • Perform Tomcat-specific connection tasks
          • Start Your Server
        • Use MS SQL Server as your repository database (Archive installation)
          • Before you begin
          • Initialize MS SQL Server Pentaho Repository database
            • Step 1: Adjust MS SQL Server configuration settings
            • Step 2: Change default passwords
            • Step 3: Run SQL scripts
            • Step 4: Verify MS SQL Server initialization
          • Configure MS SQL Server Pentaho Repository database
            • Step 1: Set up Quartz on MS SQL Server
            • Step 2: Set Hibernate settings for MS SQL Server
            • Step 3: Replace default version of audit log file with MS SQL Server version
            • Step 4: Modify Jackrabbit repository information for MS SQL Server
          • Perform Tomcat-specific connection tasks
          • Start Your Server
        • Starting the Pentaho Server after an archive installation
      • Manual installation of the Pentaho Client Tools
      • Troubleshooting
    • Manual installation
      • Before you begin
      • Manual installation process
        • Prepare your Windows environment for a manual installation
          • Process overview
          • Create Windows directory structure
          • Install Java
          • Install your web application server
          • Install the Pentaho Repository host database
          • Download and unpack installation files
            • Step 1: Download files
            • Step 2: Unpack the Pentaho Server installation files
            • Step 3: (Optional) Unpack the plugin files
            • Step 4: Verify directory structure
          • Set environment variables
          • Prepare your Pentaho Repository
        • Prepare your Linux environment for a manual installation
          • Process overview
          • Create the Pentaho user
          • Create Linux directory structure
          • Install Java
          • Install your web application server
          • Install the Pentaho Repository host database
          • Download and unpack installation files
          • Set environment variables
          • Advanced Linux considerations
          • Prepare your Pentaho Repository
        • Use PostgreSQL as your repository database (Manual installation)
          • Before you begin
          • Initialize PostgreSQL Pentaho Repository database
          • Configure PostgreSQL Pentaho Repository database
            • Step 1: Set up Quartz on PostgreSQL
            • Step 2: Set Hibernate settings for PostgreSQL
            • Step 3: Modify Jackrabbit repository information for PostgreSQL
          • Perform Tomcat-specific connection tasks
          • Next steps
        • Use MySQL or MariaDB as your repository database (Manual installation)
          • Before you begin
          • Initialize MySQL or MariaDB Pentaho Repository database
          • Configure MySQL or MariaDB Pentaho Repository Database
          • Perform Tomcat-specific connection tasks
          • Next steps
        • Use Oracle as your repository database (Manual installation)
          • Before you begin
          • Initialize Oracle Pentaho Repository database
          • Configure Oracle Pentaho Repository database
          • Perform Tomcat-specific connection tasks
          • Next steps
        • Use MS SQL Server as your repository database (Manual installation)
          • Before you begin
          • Initialize MS SQL Server Pentaho Repository database
          • Configure MS SQL Server Pentaho Repository database
          • Perform Tomcat-specific connection tasks
          • Next steps
        • Configure and start the Pentaho Server after manual installation
          • Configure the Pentaho Server for Windows
            • Step 1: Modify the Tomcat startup script
            • Step 2: (Optional) Run the Pentaho Server as a Windows service
            • Step 3: Install licenses
            • Step 4: Start the Pentaho Server
          • Configure the Pentaho Server for Linux
            • Step 1: Modify the Tomcat startup script
            • Step 2: Install licenses
            • Step 4: Start the Pentaho Server
          • Starting the Pentaho Server
          • Problems starting the Pentaho Server
      • Manual installation of the Pentaho Client Tools
    • Installation of the Pentaho design tools
      • Step 1: Download files
      • Step 2: Unpack the files
      • Step 3: Install the design tools
      • (Optional) Install the Analysis Enterprise Edition plugin
      • Install PDI tools and plugins
        • Step 1: Download files
        • Step 2: Unpack the files
        • Step 3: Install PDI
        • Step 4: Install PDI plugins
        • PDI steps and entries included in the Hadoop add-on installation
    • Docker container deployment of Pentaho
      • Before you begin
      • Docker container deployment process
        • Download and install DockMaker command line tool
        • Using the DockMaker command line tool
          • DockMaker command line tool examples
          • Docker command tool property and registry files
            • Docker command tool property file
              • Preset variables
              • Derived variables
              • Mandatory variables
              • Variable resolution order
            • Docker command tool registry file
        • Starting or stopping your Docker container
          • Starting or stopping a Pentaho Server container
          • Running a PDI container
          • Starting or stopping a Carte server container
          • Getting a command prompt on a container
      • Using your Docker containers with clusters
        • Shared volumes
        • Use the Docker command tool with a Kerberos secured cluster
      • Using DockMaker with service packs
    • Hyperscalers
      • Installing Pentaho on AWS
        • Installing the Platform Server or PDI Server on AWS
          • Prerequisites for installing the Platform or PDI Server on AWS
          • Process overview for installing the Platform or PDI Server on AWS
          • Download and extract Platform or PDI Server for AWS
          • Create an Amazon ECR
          • Load and push the Pentaho Docker image to ECR
          • Create an RDS database
          • Create an S3 bucket
          • Create an EKS cluster and add a node group
          • Install the Platform or PDI Server on AWS
          • Dynamically update server configuration content from S3
          • Worksheet for AWS hyperscaler
        • Installing the Carte Server on AWS
          • Prerequisites for installing the Carte Server on AWS
          • Process overview for installing the Carte Server on AWS
          • Download and extract Pentaho for AWS
          • Create an Amazon ECR
          • Load and push the Pentaho Docker image to ECR
          • Create an S3 bucket for the Carte Server
          • Worksheet for AWS hyperscaler
        • Running PDI-CLI on AWS
          • Prerequisites for installing PDI-CLI on AWS
          • Process overview for running PDI-CLI on AWS
          • Download and extract Pentaho for AWS
          • Create an Amazon ECR for PDI-CLI
          • Load and push the PDI-CLI Docker image to ECR
          • Load and push the Pentaho Docker image to ECR
          • Create an S3 bucket for PDI-CLI
          • Configure and execute PDI-CLI
          • Worksheet for AWS hyperscaler
    • Troubleshooting possible installation and upgrade issues
      • User Console themes render improperly after upgrade
      • Context XML changes do not take effect after deploying a WAR
      • No search manager configured for the Java repository
      • Cannot create Jackrabbit tables in MySQL
      • Unable to use the database init scripts for PostgreSQL
      • Not able to login or run transformations
      • Error when using sample transformations or jobs
      • Missing Quartz library database error
      • Error when installing local license server on Redhat Linux in FIPS mode
  • Components Reference
  • JDBC drivers reference
    • Amazon Redshift
    • AWS Athena
    • Caché
    • CUBRID
    • Daffodil DB
    • DB2 AS/400
    • DB2 Universal Database
    • Firebird
    • FrontBase
    • Google BigQuery
    • Greenplum
    • H2 Database
    • Hive
    • Hive2
    • HSQLDB
    • Impala
    • Informix
    • Ingres
    • InterBase
    • jTDS Free MS SQL Sybase
    • MariaDB
    • MaxDB
    • Mckoi SQL Database
    • Mimer
    • MySQL
    • Neoview
    • Netezza
    • OpenBase SQL
    • Oracle
    • Pervasive
    • PostgreSQL
    • SAP ASE (formerly Sybase ASE)
    • SAP DB
    • SAP HANA
    • SAP SQL Anywhere
    • SmallSQL
    • Snowflake
    • SQLite
    • SQL Server
    • Teradata
    • Vertica
    • Install drivers with the JDBC distribution tool
  • Pentaho configuration
    • Tasks to be performed by an IT administrator
      • Configure the Pentaho Server
        • Stop and start the Pentaho Server and repository
        • Create scripts for automatic stop and start of the Pentaho server and repository on Linux
        • Increase Pentaho Server memory limit for custom installations on Windows or Linux
        • Specify data connections for BA design tools
          • JDBC Database Connections
            • Define JDBC or OCI connections for BA design tools
              • Add drivers
            • Define JNDI connections for Report Designer and Metadata Editor
              • Add drivers
              • Edit the properties file for Report Designer
              • Edit the properties file for Metadata Editor
        • Specify data connections for the Pentaho Server
          • JDBC database connections
            • Set up native (JDBC) or OCI data connections for the Pentaho Server
              • Add drivers
                • Driver for Microsoft SQL Server
            • Set up JNDI connections for the Pentaho Server
              • Defining JNDI connections for PDI clients
              • Tomcat JNDI connections
                • Add drivers
                  • Driver for Microsoft SQL Server
      • Define security for the Pentaho Server
      • Use password encryption with Pentaho
        • Encrypting a password
        • Using encrypted passwords with Pentaho products
          • Encrypted passwords with Pentaho Data Integration
          • Encrypted passwords with the Pentaho User Console
          • Encrypted passwords with PUC email
          • Encrypted passwords with the Pentaho Aggregation Designer
          • Encrypted passwords with the Pentaho Metadata Editor
          • Encrypted passwords with the Pentaho Report Designer
      • Set up Pentaho to connect to a Hadoop cluster
        • Adding a new driver
          • Add a new driver
        • Install a driver for the Pentaho Server
        • Manually install a driver for the Pentaho Server
        • Additional configurations for specific distributions
          • Advanced settings for connecting to an Amazon EMR cluster
            • Before you begin
            • Edit configuration files for users
              • Verify or edit core-site XML file
              • Edit mapred-site XML file
            • Connect to a Hadoop cluster with the PDI client
            • Connect other Pentaho components to the Amazon EMR cluster
              • Create and test connections
          • Advanced settings for connecting to Azure HDInsight
            • Before you begin
              • Set up a secured instance
            • Edit configuration files for users
              • Edit Core site XML file
              • Edit HBase site XML file
              • Edit Hive site XML file
              • Edit Mapred site XML file
              • Edit YARN site XML file
            • Oozie configuration
              • Set up Oozie on a cluster
              • Set up Oozie on the server
            • Windows configuration for a secured cluster
            • Connect to HDI with the PDI client
            • Connect other Pentaho components to HDI
              • Create and test connections for other Pentaho components
          • Advanced settings for connecting to Cloudera Data Platform
            • Before you begin
              • Set up a secured instance of CDP
            • Edit configuration files for users
              • Edit Core site XML file
              • Edit Hive site XML file
              • Edit Mapred site XML file
              • Edit YARN site XML file
            • Oozie configuration
            • Windows configuration for a secured cluster
            • Connect to CDP with the PDI client
            • Connect other Pentaho components to CDP
              • Create and test connections for other Pentaho components
          • Advanced settings for connecting to Google Dataproc
            • Before you begin
            • Create a Dataproc cluster
            • Install the Google Cloud SDK on your local machine
              • Set command variables
            • Set up a Google Compute Engine instance for PDI
            • Edit configuration files for users
              • Edit the XML file for MapReduce
            • Connect to a Hadoop cluster with the PDI client
            • Connect other Pentaho components to Dataproc
              • Create and test connections
    • Tasks to be performed by a Pentaho administrator
      • Define data connections
        • Open the connection dialog box
          • Open the Database Connection dialog box from PUC
          • Open the Database Connection dialog box from PDI
        • Enter database connection information
          • Native (JDBC) protocol information
          • OCI protocol information (PDI only)
          • Connect to Snowflake using strong authentication
          • Connect to an Azure SQL database
            • Before you begin
            • Authentication method
            • Connect to an Azure database
            • Use the Always Encryption Enabled option
        • Edit existing connections
          • Edit database connections in PUC
          • Edit database connections in PDI
        • Specify advanced configuration of database connections
        • Quoting
        • Set database-specific options
        • Define connection pooling
        • Connect to clusters (PDI only)
        • Modify connections
          • Modify connections from PUC
          • Modify connections from PDI
        • Delete connections
          • Delete connections from PUC
          • Delete connections from PDI
      • Assign permissions to use or manage database connections
      • Manage users and roles
        • With PUC
        • With the PDI client
      • Configure the design tools and utilities
        • BA design tools
          • Start and stop BA design tools
            • Starting BA design tools
            • Stopping BA design tools
        • PDI design tools and utilities
          • Start and stop PDI design tools and utilities
          • Increase the PDI client memory limit
  • Pentaho upgrade
    • Before you begin
    • Download the Pentaho upgrade file
    • Using the Pentaho Upgrade Installer
      • Use the Pentaho Upgrade Installer to upgrade archive installation
        • Get started by checking your environment
        • Specify customized items to address after upgrading
        • Back up your existing Pentaho products and install Pentaho 10.2
      • Use the Pentaho Upgrade Installer to upgrade manual installation
        • Install upgrade files into empty folder
        • Update files in manual installation
      • Restoring a Pentaho Upgrade Installer backup of archive installation
        • Use the Pentaho Upgrade Installer to restore a backup of archive installation
    • Mandatory Quartz upgrade for versions 10.2.0.1 and later
      • Initialize Quartz for PostgreSQL
      • Initialize Quartz for MySQL or MariaDB
      • Initialize Quartz for Oracle
      • Initialize Quartz for MS SQL Server
    • Post-upgrade tasks
      • Apply customizations
        • Address customizations to upgraded archive installation
        • Address customizations to upgraded manual installation
      • Update the default documentation version link
      • Install Ops Mart
      • Install drivers for your Hadoop clusters
      • Apply your plugins
      • Setting up password encryption after upgrading
        • Modify the Tomcat context XML file
        • Update the Jackrabbit Repository XML file
        • Verify your Quartz properties
        • Update your Hibernate configuration
    • Using the Pentaho Upgrade Installer in console or silent mode
      • Before you begin
      • Using silent mode to upgrade Pentaho products
        • Use silent mode to upgrade archive installation
        • Use silent mode to upgrade manual installation
      • Exit codes
    • Service pack information
  • Multidimensional Data Modeling in Pentaho
    • Prepare your data
    • Dimensional modeling
    • Understanding data cubes
    • Map a model with Schema Workbench
    • Configure Mondrian engine
    • Mondrian cache control
      • Switch to Memcached
        • Switch from the Infinispan cache framework
          • Memcached Configuration Options
      • Segment cache architecture
        • How the Analysis engine uses memory
        • Cache control and propagation
      • Cache Configuration Files
      • Modify the JGroups configuration
      • Switch to another cache framework
        • Switch to Pentaho Platform Delegating Cache
        • Use a Custom SegmentCache SPI
    • Analysis schema security
      • Restrict Access to Specific Members
      • Mondrian role mapping in the Pentaho Server
        • The Mondrian One-To-One UseRoleMapper
        • The Mondrian-SampleLookupMap-UserRoleMapper
        • The Mondrian-SampleUserSession-UserRoleMapper
    • OLAP Log Output
      • Enabling segment cache logging
      • View Log Output in Analyzer
    • Multidimensional Expression Language
      • Mondrian Schema Element Reference
        • AggExclude
          • Attributes
          • Constituent Elements
        • AggFactCount
          • Attributes
          • Constituent Elements
        • AggForeignKey
          • Attributes
          • Constituent Elements
        • AggIgnoreColumn
          • Attributes
          • Constituent Elements
        • AggLevel
          • Attributes
          • Constituent Elements
        • AggMeasure
          • Attributes
          • Constituent Elements
        • AggName
          • Attributes
          • Constituent Elements
        • AggPattern
          • Attributes
          • Constituent Elements
        • AggTable
          • Attributes
          • Constituent Elements
        • CalculatedMember
          • Attributes
          • Constituent Elements
        • CalculatedMemberProperty
          • Attributes
          • Constituent Elements
        • CaptionExpression
          • Attributes
          • Constituent Elements
        • Closure
          • Attributes
          • Constituent Elements
        • ColumnDef
          • Attributes
          • Constituent Elements
        • ColumnDefs
          • Attributes
          • Constituent Elements
        • Cube
          • Attributes
          • Constituent Elements
        • CubeGrant
          • Attributes
          • Constituent Elements
        • CubeUsage
          • Attributes
          • Constituent Elements
        • CubeUsages
          • Attributes
          • Constituent Elements
        • Dimension
          • Attributes
          • Constituent Elements
        • DimensionGrant
          • Attributes
          • Constituent Elements
        • DimensionUsage
          • Attributes
          • Constituent Elements
        • Formula
          • Attributes
          • Constituent Elements
        • Hierarchy
          • Attributes
          • Constituent Elements
        • HierarchyGrant
          • Attributes
          • Constituent Elements
        • InlineTable
          • Attributes
          • Constituent Elements
        • Join
          • Attributes
          • Constituent Elements
        • KeyExpression
          • Attributes
          • Constituent Elements
        • Level
          • Attributes
          • Constituent Elements
        • Measure
          • Attributes
          • Constituent Elements
        • MeasureExpression
          • Attributes
          • Constituent Elements
        • MemberGrant
          • Attributes
          • Constituent Elements
        • NamedSet
          • Attributes
          • Constituent Elements
        • NameExpression
          • Attributes
          • Constituent Elements
        • OrdinalExpression
          • Attributes
          • Constituent Elements
        • Parameter
          • Attributes
          • Constituent Elements
        • ParentExpression
          • Attributes
          • Constituent Elements
        • Property
          • Attributes
          • Constituent Elements
        • PropertyExpression
          • Attributes
          • Constituent Elements
        • Role
          • Attributes
          • Constituent Elements
        • RoleUsage
          • Attributes
          • Constituent Elements
        • Row
          • Attributes
          • Constituent Elements
        • Rows
          • Attributes
          • Constituent Elements
        • Schema
          • Attributes
          • Constituent Elements
        • SchemaGrant
          • Attributes
          • Constituent Elements
        • SQL
          • Attributes
          • Constituent Elements
        • Table
          • Attributes
          • Constituent Elements
        • Union
          • Attributes
          • Constituent Elements
        • UserDefinedFunction
          • Attributes
          • Constituent Elements
        • Value
          • Attributes
          • Constituent Elements
        • View
          • Attributes
          • Constituent Elements
        • VirtualCube
          • Attributes
          • Constituent Elements
        • VirtualCubeDimension
          • Attributes
          • Constituent Elements
        • VirtualCubeMeasure
          • Attributes
          • Constituent Elements
  • Relational Data Modeling in Pentaho
    • The Physical Layer
    • The Business View
    • The Abstract Business Layer
    • Incorporate Metadata
  • Use Hadoop with Pentaho
    • Get started with Hadoop and PDI
      • Before you begin
        • Configure PDI for Hadoop connections
          • Include or exclude classes or packages for a Hadoop configuration
        • Hadoop connection and access information list
      • Connect to your Hadoop clusters in the PDI client
      • Use PDI outside and inside the Hadoop cluster
    • Advanced topics
      • Copy files to a Hadoop YARN cluster
        • Add files to the YARN Workspace folder
        • Delete files from the YARN Workspace folder
      • PDI big data transformation steps
      • PDI big data job entries
      • Big data resources
    • Troubleshooting possible Big Data issues
      • Cannot access cluster with Kerberos enabled
      • Cannot access the Hive service on a cluster
      • HBase Get Master Failed error
      • Sqoop export fails
      • Sqoop import into Hive fails
      • Group By step is not supported in a single threaded transformation engine
      • Kettle cluster on YARN will not start
      • Hadoop on Windows
      • Legacy mode activated when named cluster configuration cannot be located
      • Unable to read or write files to HDFS on the Amazon EMR cluster
      • Use YARN with S3
      • Data Catalog searches returning incomplete or missing data
  • Using Spark Submit
    • Modify the sample Spark Submit job
Powered by GitBook
On this page

Was this helpful?

Export as PDF
  1. Pentaho installation
  2. Hyperscalers
  3. Installing Pentaho on AWS
  4. Installing the Platform Server or PDI Server on AWS

Worksheet for AWS hyperscaler

To access the common worksheet for the AWS hyperscaler, go to Worksheet for AWS hyperscaler.

PreviousDynamically update server configuration content from S3NextInstalling the Carte Server on AWS

Last updated 25 days ago

Was this helpful?

LogoLogo

About

  • Pentaho.com

Support

  • Pentaho Support

Resources

  • Privacy

© 2025 Hitachi Vantara LLC