Performance monitoring
Use logging and data marts to monitor Pentaho performance.
Use PDI logging for transformation and job status.
Use Pentaho Server logging for exceptions and debugging.
Use Pentaho Operations Mart for prebuilt performance and audit reports.
Use SNMP to integrate third-party monitoring tools.
Best practices for logging
Kettle logging lets you control location, detail level, and output format.
Store logs in a centralized database.
Ensure the logging user can insert into log tables.
Install JDBC drivers on every server and client.
Use implied schemas when possible.
Use templates for jobs and transformations with logging enabled.
Use global Kettle logging variables when possible.
Use separate log tables for jobs and transformations.
If you use kettle.properties:
Back up
kettle.propertiesbefore making bulk edits.Keep a master copy and sync it across nodes.
Test changes locally, then on the server.
Tracking access to sensitive data
You can log who accessed sensitive data and when.
This helps with audits and incident response.
You can correlate users and IP addresses with SQL, MDX, and parameter values.
Supported tools:
Analyzer
Mondrian
Interactive Reports (PRTI)
Dashboards
Report Designer (PRPT)
CTools (CDA)
Content run by the Pentaho Scheduler
MDC tracking properties
These properties use Log4j2 Mapped Diagnostic Context (MDC).
Add the properties you want to log4j2.xml.
sessionId: Pentaho session ID. Usually the JSession cookie.sessionName: User name.instanceId: UUID for a report execution instance.remoteAddr: Client IP address.remoteHost: Client host.remotePort: Client port.serverName: Server name accessed by the client.serverPort: Server port accessed by the client.localAddr: Local server IP.localName: Local server name.localPort: Local server port.
Pentaho writes a rolling session_data_access.log file.
Enable MDC logging
Work with Operations Mart
MDC properties work well with the BA Operations Mart.
You can join MDC attributes to the PRO_AUDIT table:
sessionId→PRO_AUDIT.INST_IDsessionName→PRO_AUDIT.ACTORinstanceId→PRO_AUDIT.MESSAGE_NAME
Correlate logs with third-party tools
For large environments, index logs with tools like Splunk or an ELK stack.
Maintain logging
You can maintain system logs using rotation and by monitoring execution status.
Log rotation
This procedure assumes you are not using OS-level log rotation.
Pentaho Server uses Log4j2.
The default log4j2.xml config rotates logs daily.
To rotate by file size, use SizeBasedTriggeringPolicy.
Stop all relevant servers.
Edit
server/pentaho-server/tomcat/webapps/pentaho/WEB-INF/classes/log4j2.xml.Update the
RollingFileappender. Example:Start servers and validate rotation.
Execution status
Use the PDI Status page to view scheduled and remotely executed jobs and transformations.
Open:
You must sign in first.
The list clears when the server restarts.
It also clears based on object_timeout_minutes.
On Carte (Pentaho Server embedded Carte)
Carte settings are in:
pentaho-server/pentaho-solutions/system/kettle/slave-server-config.xml
Stop the Pentaho Server before editing slave-server-config.xml.
Key settings:
max_log_lines:0means no limit.max_log_timeout_minutes:0means no timeout.object_timeout_minutes:0means no timeout.
Example:
PDI logging
PDI logging gives you job and transformation logs without digging through full server logs.
For Pentaho Server, this logging is separate from the platform pentaho.log.
Configure the PDI log file
Transformation and job logging is enabled by default.
Logging levels and rotation are configured separately for the PDI client and server.
Stop the Pentaho Server or close the PDI client.
Edit
log4j2.xml:Pentaho Server:
server/pentaho-server/tomcat/webapps/pentaho/WEB-INF/classes/log4j2.xmlPDI client:
design-tools/data-integration/classes/log4j2.xml
Set logger levels for:
org.pentaho.di.trans.Transorg.pentaho.di.job.Job
PDI log level mapping:
BASIC
INFO
DETAILED
INFO
MINIMAL
WARN
DEBUG
DEBUG
ERROR
ERROR
ROWLEVEL
TRACE
Set a rollover pattern for
pdi-execution-appenderusingFileNamePattern:
yyyy-MM-dd
Daily (default)
yyyy-MM
Monthly
yyyy-MM-dd-HH-mm
Every minute
Set the timestamp format for
pdi-execution-appenderusingPattern.
Examples:
{yyyy-MM-dd HH:mm:ss.SSS}
2012-11-02 14:34:02.123
{HH:mm:ss,SSS}
14:34:02,123
{dd MMM yyyy HH:mm:ss,SSS}
02 Nov 2012 14:34:02,123
{MMM dd,yyyy HH:mm:ss}
Nov 02,2012 14:34:02
If the format is invalid, PDI falls back to yyyy/MM/dd HH:mm:ss.
Optional: add a time zone:
{yyyy-MM-dd HH:mm:ss.SSS}{GMT-5}{HH:mm:ss}{GMT+0}
Save and restart.
Read the PDI log file
Log file locations:
Pentaho Server:
server/pentaho-server/logs/pdi.logPDI client:
design-tools/data-integration/logs/pdi.log
Common fields:
DateTimeStamp: date and time.LogThreshold: INFO, ERROR, DEBUG, WARN, TRACE.ThreadID: unique execution key.Filepath: absolute path to the transformation or job.Message: log message.
Example:
2018-03-07 11:40:36.290 INFO <Launch transformation UUID: 1246b616-a845-4cbc-9f4c-8a4a2cbfb4f1> [...] Starting entry
Third-party monitoring with SNMP
You can use the SNMP plugin to integrate third-party tools for monitoring PDI events.
These events help you identify long-running jobs and transformations.
Events can also capture milestones like database connections and Carte startup.
Before you begin
Download pentaho-monitoring-plugin-mib-assembly-11.0.0.0-<build number>.zip from the Support Portal.
The ZIP contains the MIB file you upload to your monitoring tool.
Send traps through a centralized monitoring server
You can use a Pentaho Server as a JMS monitoring server.
That server forwards events to your SNMP manager.
Step 1: Configure Pentaho Server to forward traps
You need the IP address of the JMS monitoring server.
Stop the Pentaho Server.
Open the SNMP config file in:
pentaho-solutions/system/karaf/system/pentaho/pentaho-osgi-config/11.0.0.0-<build number>/Update:
fromHost: IP of this Pentaho ServertoHost: IP of the JMS monitoring serverport (default is 162)
In
pentaho-solutions/system/karaf/etc/org.apache.karaf.features.cfg, add:If
pentaho-server/pentaho-solutions/system/karaf/etc/pentaho.snmp.cfgexists, delete it.Start the Pentaho Server.
Step 2: Configure your monitoring tool
Steps vary by tool.
For Nagios-style workflows:
Upload
PENTAHO-MIB-V2C.mibto your monitoring tool.Convert it to a
.conffile usingsnmpttconvertmib.Add the generated
.conffile tosnmptt.ini.Restart the SNMP components and the monitoring tool.
Send traps directly from PDI tools
You can also send traps from PDI tools like Pan, Kitchen, Spoon, or Carte.
Enable the pentaho-monitoring-to-snmp feature in the relevant Karaf features file.
Configure monitored event types
By default, Pentaho monitors all supported event types.
To reduce noise, edit the comma-separated list in monitoring.properties.
Keep only the event types you want.
Restart the Pentaho Server after changes.
Available extension points
These extension points can emit monitoring events:
TransformationPrepareExecution
Transformation prepares execution
TransformationStart
Transformation starts
TransformationHeartbeat
Periodic signal during execution
TransformationFinish
Transformation finishes
JobStart
Job starts
JobHeartbeat
Periodic signal during execution
JobFinish
Job finishes
JobBeforeJobEntryExecution
Before a job entry runs
JobAfterJobEntryExecution
After a job entry runs
DatabaseConnected
After a successful connection
DatabaseDisconnected
After a successful disconnection
CarteStartup
After Carte starts
CarteShutdown
Before Carte stops
Pentaho Operations Mart
Pentaho Operations Mart helps you monitor Pentaho Server performance.
It includes:
Business Analytics (BA) Operations Mart. Aggregates Pentaho Server audit data.
Data Integration (DI) Operations Mart. Stores and summarizes PDI logging data.
Business Analytics Operations Mart
Pentaho can automatically set up Operations Mart content for Analyzer, Interactive Reports, Report Designer, and dashboards.
BA Operations Mart aggregates data from Pentaho Server log files into prebuilt audit reports.
If the prebuilt reports do not meet your needs, use the DI Operations Mart to change them.
Download and install Operations Mart files
Operations Mart files are stored in a prepackaged ZIP file.
To install, stop the Pentaho Server, download and unpack the ZIP, then restart the server to import the files.
Stop the Pentaho Server.
See the Install Pentaho Data Integration and Analytics document for start and stop steps.
Download
pentaho-operations-mart-11.0.0.0-<build number>.zipfrom the Support Portal.Sign in using the support credentials from your welcome packet.
In the Pentaho card, select Download.
In the 11.x list, select Pentaho 11.0 GA Release.
Scroll to the bottom of the release page.
In the file component section, select the
Operations Martfolder.Download the ZIP file.
Unpack the ZIP file to a temporary directory.
Move the unpacked files to
pentaho/server/pentaho-server/pentaho-solutions/system/default-content.In
default-content, delete files you do not need for your repository database type.Repository database typeFiles to keepPostgreSQL
pentaho-operations-mart-clean-11.0.0.0-<build number>.zip*pentaho-operations-mart-etl-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zipMySQL
pentaho-operations-mart-clean-mysql5-11.0.0.0-<build number>.zip*pentaho-operations-mart-etl-mysql5-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zipOracle
pentaho-operations-mart-clean-oracle10g-11.0.0.0-<build number>.zip*pentaho-operations-mart-etl-oracle10g-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zipMicrosoft SQL Server
pentaho-operations-mart-clean-mssql-11.0.0.0-<build number>.zip*pentaho-operations-mart-etl-mssql-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zip* Keep the
pentaho-operations-mart-clean-<database>-11.0.0.0-<build number>.zipfile only if you want Pentaho to delete old entries on a schedule. See Clean up the BA Operations Mart.Restart the Pentaho Server.
Increase the maximum character length in audit table fields
You can increase the maximum audit table field length from 200 to 1024 characters.
The scripts in this procedure adjust the column width and reindex the Operations Mart tables.
Stop the Pentaho Server.
See the Install Pentaho Data Integration and Analytics document for start and stop steps.
Download
pentaho-server-ee-11.0.0.0-<build number>.zipfrom the Support Portal.Unpack the ZIP to a temporary directory.
Go to
pentaho/server/pentaho-server/data.Open the directory for your repository database type:
mysql,oracle,postgresql, orsqlserver.Locate:
alter_script_<repository database type>_BISERVER-13674.sqlpentaho_mart_upgrade_audit_<repository database type>.sql
Run the scripts in order.
Delete the temporary directory.
Start the Pentaho Server.
Choose a prebuilt Operations Mart report
Choose the report that fits your needs.
Then follow View and edit Operations Mart reports.
Time to run a report
Content Duration
Content failures by time range
Content Failures
Compare login metrics by day in month
Content Request Day of Month
Compare login metrics by day in week
Day of Week Request and Login Metrics
Content sorted by type
Content Type Usage
Content usage by time range
Content Usage
Compare login metrics by hour
Hours in Day Request and Login Metrics
Login duration and logins per user
Session Duration Counts
View and edit Operations Mart reports
Pick a report from Choose a prebuilt Operations Mart report.
Double-click Browse Files.
Select public > Pentaho Operations Mart > BA Audit Reports.
Select the file, then click Open.
Optional: edit the report in the tool you used to open it.
Create Operations Mart reports
If the prebuilt reports do not meet your needs, create your own.
In the User Console, select Create New.
Select the report type.
Select the data source you need.
Information shownData source.xactionexecution detailspentaho_operations_mart: BA Operations Mart - ComponentContent execution details
pentaho_operations_mart: BA Operations Mart - ContentSession metrics
pentaho_operations_mart: BA Operations Mart - User SessionJob entry executions
pentaho_operations_mart: PDI Operations Mart - Job EntryServer performance detail
pentaho_operations_mart: - PDI Operations Mart - PerformanceStep executions
pentaho_operations_mart: PDI Operations Mart - StepTransformation and job runs
pentaho_operations_mart: PDI_Operations_Mart
Update the Operations Mart
In the User Console, select Browse Files > public > Pentaho Operations Mart > Update Audit Mart.
Double-click Update Operations Mart Date & Time or Update BA Operations Mart Data.
View a report to confirm updates.
Clean up the BA Operations Mart
Operations Mart tables remove entries older than 365 days by default.
To change the age, add ba.cleanup.max.age to {user.home}/.kettle/kettle.properties.
Stop the Pentaho Server.
Open
{user.home}/.kettle/kettle.properties.Add
ba.cleanup.max.age. Example:Save the file.
Restart the Pentaho Server.
Data Integration Operations Mart
The DI Operations Mart stores and summarizes PDI log data.
Use it to monitor jobs and transformations, and build reports and dashboards.
Set up the Data Integration Operations Mart
Setup depends on your installation method.
Follow these instructions if you installed Pentaho using the archive method.
Required databases
Your Pentaho Server and repository must use one of:
PostgreSQL
MySQL or MariaDB
Oracle
Microsoft SQL Server
Required scripts
You need:
pentaho_logging_<databasename>.sqlpentaho_mart_<databasename>.sql
<databasename> matches your repository database type:
postgresqlmysql15(MariaDB only)oracle10gororacle12csqlserver
Overview
Step 1: Get the Data Integration Operations Mart files
Step 2: Run the setup scripts
Step 3: Set the global Kettle logging variables
Step 4: Add the JNDI connections for logging
Step 5: Add a JDBC connection for the Pentaho Server
Step 6: Add the DI Operations Mart ETL solutions to the default content folder
Step 7: Initialize the DI Operations Mart
Step 8: Verify the DI Operations Mart is working
Step 1: Get the Data Integration Operations Mart files
If you used a PostgreSQL repository, skip to Step 2.
If you do not have pentaho-operations-mart-11.0.0.0-<build number>.zip, download it from the Support Portal.
Download
pentaho-server-ee-11.0.0.0-<build number>.zipfrom the Support Portal.Unpack the ZIP file to a temporary directory.
Verify these files exist:
pentaho-operations-mart-operations-di-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zip
Verify you also have the two files required for your repository database type:
Repository database typeRequired filesOracle
pentaho-operations-mart-etl-oracle10g-11.0.0.0-<build number>.zippentaho-operations-mart-clean-oracle10g-11.0.0.0-<build number>.zipMySQL5
pentaho-operations-mart-etl-mysql5-11.0.0.0-<build number>.zippentaho-operations-mart-clean-mysql5-11.0.0.0-<build number>.zipMicrosoft SQL Server
pentaho-operations-mart-etl-mssql-11.0.0.0-<build number>.zippentaho-operations-mart-clean-mssql-11.0.0.0-<build number>.zip
Step 2: Run the setup scripts
Scripts are in:
<install-directory>/pentaho-server/data/<databasename>
Run in order:
pentaho_logging_<databasename>.sqlpentaho_mart_<databasename>.sql
Step 3: Set the global Kettle logging variables
Run this on the machine with the PDI client and Pentaho Server.
kettle.properties is created at:
$USER_HOME/.kettle/kettle.properties
In Spoon, select Edit > Edit the kettle.properties file.
Add or edit variables for the log tables.
For Oracle and Microsoft SQL Server, leave the value blank for variables that contain
_SCHEMA_.VariableValueKETTLE_CHANNEL_LOG_DBlive_logging_infoKETTLE_CHANNEL_LOG_TABLEchannel_logsKETTLE_CHANNEL_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_JOBENTRY_LOG_DBlive_logging_infoKETTLE_JOBENTRY_LOG_TABLEjobentry_logsKETTLE_JOBENTRY_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_JOB_LOG_DBlive_logging_infoKETTLE_JOB_LOG_TABLEjob_logsKETTLE_JOB_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_METRICS_LOG_DBlive_logging_infoKETTLE_METRICS_LOG_TABLEmetrics_logsKETTLE_METRICS_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_STEP_LOG_DBlive_logging_infoKETTLE_STEP_LOG_TABLEstep_logsKETTLE_STEP_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_TRANS_LOG_DBlive_logging_infoKETTLE_TRANS_LOG_TABLEtrans_logsKETTLE_TRANS_LOG_SCHEMApentaho_dilogsVariableValueKETTLE_TRANS_PERFORMANCE_LOG_DBlive_logging_infoKETTLE_TRANS_PERFORMANCE_LOG_TABLEtransperf_logsKETTLE_TRANS_PERFORMANCE_LOG_SCHEMApentaho_dilogs
Step 4: Add the JNDI connections for logging
Go to
<pdi-install>/data-integration/simple-jndi.Open
jdbc.properties.Update properties for your database type.
PostgreSQL example:
Step 5: Add a JDBC connection for the Pentaho Server
Go to
<pentaho-server>/tomcat/webapps/pentaho/META-INF/.Open
context.xml.Update the resources for your database type.
Optional: for encrypted passwords, set:
factory="org.pentaho.di.core.database.util.DecryptingDataSourceFactory"
Step 6: Add the DI Operations Mart ETL solutions to the default content folder
If you use PostgreSQL for the repository database, skip to Step 7.
Stop the Pentaho Server.
Locate:
pentaho-operations-mart-operations-di-11.0.0.0-<build number>.zippentaho-operations-mart-operations-bi-11.0.0.0-<build number>.zip
Locate the two ZIP files that match your repository database type.
Copy all four ZIP files into:
$PENTAHO_HOME/pentaho-server/pentaho-solution/system/default-contentStart the Pentaho Server.
After processing, the server renames the ZIP files with a timestamp. Keep the renamed files.
Step 7: Initialize the DI Operations Mart
Launch Spoon.
Connect to the Pentaho Repository through the Pentaho Server.
Select File > Open.
Select Browse Files > Public > Pentaho Operations Mart > DI Ops Mart ETL.
Open each transformation and job.
In each, open Properties > Logging.
Run
Fill_in_DIM_DATE_and_DIM_TIME.Run a few sample KTRs to generate log activity.
Run
Update_Dimensions_then_Logging_Data.
Step 8: Verify the DI Operations Mart is working
In the User Console, select Browse Files > Public > Pentaho Operations Mart > DI Audit Reports > Last_Run.
Verify jobs and transformations ran.
Follow these instructions if you installed Pentaho using the manual method.
Steps match the archive method, except you first unpack the DDL ZIP.
Download
pentaho-operations-mart-11.0.0.0-<build number>.zip.Unpack it to a temporary directory.
Locate
pentaho-operations-mart-ddl-11.0.0.0-<build number>.zip.Move it to
<pentaho-server>/data/and unpack it.
This writes scripts into the repository database directory.
Then follow the archive steps for scripts, variables, connections, content, and initialization.
Charts, reports, and dashboards
After you populate the DI Operations Mart, use the User Console to create reports, charts, and dashboards.
Pentaho provides prebuilt content you can modify.
To understand fields, see Logging dimensions and metrics.
Logging tables status
Transformation log status values
start
Started. Stays until end when no interval is set.
end
Ended successfully.
stop
Stopped by a user.
error
Failed with an error.
running
Briefly after starting. Not shown without an interval.
paused
Paused by a user. Not shown without an interval.
Job log status values
start
Started. Stays until end when no interval is set.
end
Ended successfully.
stop
Stopped by a user.
error
Failed with an error.
running
Briefly after starting. Not shown without an interval.
paused
Paused by a user. Not shown without an interval.
Logging dimensions and metrics
These tables identify the dimensions and metrics used to create ETL log charts and reports.
Fact table (fact_execution)
execution_date_tk
Technical key linking to the execution date.
execution_time_tk
Technical key linking to the execution time.
batch_tk
Technical key linking to batch information.
execution_tk
Technical key linking to execution information.
executor_tk
Technical key linking to executor information.
parent_executor_tk
Technical key linking to parent executor information.
root_executor_tk
Technical key linking to root executor information.
execution_timestamp
Date and time of execution.
duration
Duration in seconds (LOGDATE to max DEPDATE).
rows_input
Rows read from disk or network by the step.
rows_output
Rows output during execution.
rows_read
Rows read from the step input stream.
rows_written
Rows written during execution.
rows_rejected
Rows rejected during execution.
errors
Errors during execution.
Batch dimension (dim_batch)
batch_tk
Technical key for batch info.
batch_id
Batch ID.
logchannel_id
Logging channel ID.
parent_logchannel_id
Parent logging channel ID.
Date dimension (dim_date)
date_tk
Technical key linking to date.
date_field
Date value.
ymd
YYYY-MM-DD.
ym
YYYY-MM.
year
Year.
quarter
Quarter number (1-4).
quarter_code
Q1 to Q4.
month
Month number (1-12).
month_desc
Month name.
month_code
JAN, FEB, and so on.
day
Day number (1-31).
day_of_year
Day of year (1-366).
day_of_week
Day of week (1-7).
day_of_week_desc
Day name.
day_of_week_code
SUN, MON, and so on.
week
Week of year (1-53).
Execution dimension (dim_execution)
execution_tk
Technical key for execution info.
execution_id
Unique execution ID.
server_name
Server name.
server_host
Server host.
executing_user
User who initiated execution.
execution_status
start, stop, end, error.
Executor dimension (dim_executor)
executor_tk
Technical key for executor.
version
Executor version.
date_from
Start of validity range.
date_to
End of validity range.
executor_id
Executor ID.
executor_source
Source location.
executor_environment
Reserved for future use.
executor_type
job or transformation.
executor_name
Executor name.
executor_desc
Description.
executor_revision
Revision string.
executor_version_label
Change comments.
exec_enabled_table_logging
Y or N.
exec_enabled_detailed_logging
Y or N.
exec_enabled_perf_logging
Y or N.
exec_enabled_history_logging
Y or N.
last_updated_date
Last update date.
last_updated_user
Last update user.
Log table (dim_log_table)
Do not modify dim_log_table. Use it as reference only.
log_table_tk
Technical key.
object_type
Object type.
table_connection_name
Log table connection name.
table_name
Log table name.
schema_name
Log table schema.
step_entry_table_conn_name
Step/job entry table connection name.
step_entry_table_name
Step/job entry table name.
step_entry_schema_name
Step/job entry table schema.
perf_table_conn_name
Performance table connection name.
perf_table_name
Performance table name.
perf_schema_name
Performance table schema.
Time-of-day dimension (dim_time)
time_tk
Technical key linking to time-of-day.
hms
HH:MM:SS.
hm
HH:MM.
ampm
am or pm.
hour
0-23.
hour12
1-12.
minute
0-59.
second
0-59.
Step fact table (fact_step_execution)
execution_date_tk
Technical key for execution date.
execution_time_tk
Technical key for execution time.
batch_tk
Technical key for batch.
executor_tk
Technical key for executor.
parent_executor_tk
Technical key for parent executor.
root_executor_tk
Technical key for root executor.
execution_timestamp
Execution date and time.
step_tk
Technical key for step.
step_copy
Step copy number.
rows_input
Rows read by the step.
rows_output
Rows written by the step.
rows_read
Rows read from previous steps.
rows_written
Rows written to following steps.
rows_rejected
Rows rejected by error handling.
errors
Errors during execution.
Step dimension (dim_step)
step_tk
Technical key for step/job entry.
step_id
Step/job entry name.
original_step_name
Reserved for future use.
Job entry fact table (fact_jobentry_execution)
execution_date_tk
Technical key for date.
execution_time_tk
Technical key for time.
batch_tk
Technical key for batch.
executor_tk
Technical key for executor.
parent_executor_tk
Technical key for parent executor.
root_executor_tk
Technical key for root executor.
step_tk
Technical key for job entry.
execution_timestamp
Execution timestamp.
rows_input
Rows read.
rows_output
Rows written.
rows_read
Rows read from input stream.
rows_written
Rows written to output stream.
rows_rejected
Rows rejected.
errors
Errors.
result
Y or N.
nr_result_rows
Result rows after execution.
nr_result_files
Result files after execution.
Execution performance fact table (fact_perf_execution)
execution_date_tk
Technical key for date.
execution_time_tk
Technical key for time.
batch_tk
Technical key for batch.
executor_tk
Technical key for executor.
parent_executor_tk
Technical key for parent executor.
root_executor_tk
Technical key for root executor.
step_tk
Technical key for step.
seq_nr
Snapshot sequence number.
step_copy
Step copy number.
execution_timestamp
Execution timestamp.
rows_input
Rows read during interval.
rows_output
Rows written during interval.
rows_read
Rows read from previous steps.
rows_written
Rows written to following steps.
rows_rejected
Rows rejected during interval.
errors
Errors.
input_buffer_rows
Input buffer size at snapshot time.
output_buffer_rows
Output buffer size at snapshot time.
Clean up the DI Operations Mart tables
Run a cleanup job or transformation to delete data older than a maximum age.
In Spoon, open:
Clean_up_PDI_Operations_Mart.kjb(job) orClean_up_PDI_Operations_Mart_fact_table.ktr(transformation)
Set parameters:
max.age.days(required): maximum data age in daysschema.prefix(optional): for PostgreSQL, set<schema>.(with trailing period)
Run it.
See the Pentaho Data Integration document for scheduling cleanup.
Last updated
Was this helpful?

