Troubleshooting Pentaho Data Catalog

The Pentaho Data Catalog log files contain information that can help you determine the root cause of error messages you might see. Refer to the following topics for information on how to resolve the issues causing the error messages.

Low disk space message

If you see a Low disk space message from Pentaho Data Catalog while loading images into the Docker repository, you can resolve this issue by linking the Docker root directory to another directory.

Important: The other directory should have at least 100 GB of free space.

Use the following steps to resolve this issue:

  1. Enter the following commands to link the /var/lib/docker directory to a directory with at least 100 GB of free space.

    Note: In this example, the directory with at least 100 GB of free space is <dir with min 100 GB free>. You should replace <dir with min 100 GB free> in the command with the full path to your directory with a minimum of 100 GB of free space.

    sudo systemctl stop docker
    sudo mv /var/lib/docker <dir with min 100 GB free>
    sudo ln -s <dir with min 100 GB free> /var/lib/docker
    sudo systemctl start docker
  2. Repeat the action that produced the Low disk space message.

The action should succeed without producing a Low disk space message.

Authentication failure after upgrading Remote Worker from 10.2.7 to 10.2.9

When upgrading the Remote Worker from version 10.2.7 to 10.2.9, the Remote Worker container fails to start. The startup log displays an error indicating that authentication failed with the SASL mechanism SCRAM-SHA-512.

This issue occurs because the Kafka user credentials used by the Remote Worker become invalid during the upgrade.

After the upgrade, the Remote Worker container (pdc-ws-remote) fails to start and shows an authentication error similar to the following in the log:

Workaround

You can fix this issue by resetting the Kafka SCRAM-SHA-512 password for the pdcuser on the PDC main server and restarting the Remote Worker.

Use the following steps to resolve this issue:

  1. Log in to the Kafka container on the PDC main server:

  2. Run the following command to reset the SCRAM-SHA-512 password for the Kafka user pdcuser:

  3. Exit the Kafka container.

  4. Restart the Remote Worker service:

After resetting the Kafka password and restarting the Remote Worker, authentication succeeds and the Remote Worker starts successfully.

Unable to connect to OpenSearch using HTTPS (Security plugin not initialized)

When accessing OpenSearch over HTTPS, the system may fail to connect because the OpenSearch Security plugin is enabled but not yet initialized. This occurs when the .opendistro_security index does not exist, preventing OpenSearch from recognizing user credentials, roles, TLS settings, and other security configurations.

You may see the following error in the logs:

This issue typically appears during an upgrade (for example, from PDC 10.2.1 to 10.2.6), not in fresh installations of Data Catalog 10.2.5 or later, where the security index is initialized by default.

Workaround

Perform the following steps to resolve the issue:

  1. Log in to the deployment server where Data Catalog is running.

  2. Stop all Pentaho Data Catalog containers:

./pdc.sh stop

  1. Identify the OpenSearch container IDs.

  2. Remove the OpenSearch containers.

  3. List all Docker volumes related to OpenSearch to confirm their presence:

    Typical volumes include:

  1. Delete the pdc_opensearch_data volume:

  2. Delete the pdc_opensearch_snapshots volume:

  3. Restart Data Catalog services:

  4. Verify that all the services are up and running now.

After removing the existing OpenSearch volumes and restart the system, the .opendistro_security index is reinitialized. OpenSearch initializes the Security plugin, loads its configuration successfully, and connects over HTTPS without errors.

Last updated