PDI job entries
Entries extend and expand the functionality of Pentaho Data Integration (PDI) entries. You can use the following steps in PDI.
Entries: A - F
Bulk loading
Bulk loads files located in S3 buckets into an Amazon Redshift database.
Bulk loading
Load data into an Azure SQL database from Azure Data Lake Storage.
Conditions
Check if one or several columns exist in a table on a specified connection.
File management
Convert file content between Windows and Unix. Converting to Unix will replace CRLF (carriage return and line feed) by LF (line feed).
File management
Copy or move result filenames (since version 5.0, this job entry has been renamed to Process result filenames and it handles Delete as well).
File encryption
Decrypt files encrypted with PGP (Pretty Good Privacy). This job entry needs GnuPG to work properly.
File management
Delete specified folders. If a the folder contains files, PDI will delete them all.
File encryption
Encrypt files with PGP (Pretty Good Privacy). This job entry needs GnuPG to work properly.
Conditions
Evaluate the content of a table. You can also specify an SQL query.
Entries: G - L
Entries: M - R
Pig Script Executor
Deprecated
No longer a part of the PDI distribution. Contact Pentaho Support for details.
Entries: S - Z
Big Data
Export data from the Hadoop Distributed File System (HDFS) into a relational database (RDBMS) using Apache Sqoop.
Big Data
Import data from a relational database (RDBMS) into the Hadoop Distributed File System (HDFS) using Apache Sqoop.
SSH2 Get (deprecated)
Deprecated
Get files using SSH2 (Deprecated in 5.0 in favor of the SFTP job entry).
SSH2 Put (deprecated)
Deprecated
Put files in a remote host using SSH2 (Deprecated in 5.0 in favor of the SFTP job entry).
General
Defines the starting point for job execution. Every job must have one (and only one) Start.
File encryption
Verify file signature with PGP (Pretty Good Privacy). This job entry needs GnuPG to work properly.
Last updated
Was this helpful?