Known Issues in Flow Management

Learn about the known issues in Flow Management clusters, the impact or changes to the functionality, and the workaround.

Learn about the known issues and limitations in Flow Management in this release:
KafkaRecordSink puts multiple records in one message

All the records are sent as a single Kafka message containing an array of records.

For more information, see NIFI-8326.

There is no workaround for this issue.
Kudu client preventing the creation of new tables using NiFi processors (KUDU-3297)
There is an issue in the Kudu client preventing the creation of new tables using NiFi processors. The table needs to exist before NiFi tries to push data into it. You may see this error when this issue arises:
Caused by: org.apache.kudu.client.NonRecoverableException: failed to wait for Hive Metastore notification log listener to catch up: failed to retrieve notification log events: failed to open Hive Metastore connection: SASL(-15): mechanism too weak for this user
There is no workaround for this issue.
NiFi Atlas reporting task does not work after data lake upgrade from light to medium

After you upgrade your data lake from light to medium scale, the data lake machine hostname and IP address will change. As the Atlas reporting task uses Atlas and Kafka server hostnames, after the upgrade the wrong hostnames will prevent NiFi to report into Atlas.

Update the configuration of the ReportLineageToAtlas reporting task:

  1. Open the Global menu on the NiFi UI.
  2. Click Controller settings.
  3. Select the Reporting tasks tab in the dialog box.
  4. Stop the ReportLineageToAtlas reporting task and update the configuration:
    • Replace the hostname value in the Atlas Urls configuration with the new Atlas hostname.
    • Replace the hostnames value in the Kafka Bootstrap servers configuration with the new Kafka bootstrap server hostnames.
  5. Start the ReportLineageToAtlas reporting task.

Technical Service Bulletins

TSB 2022-580: NiFi Processors cannot write to content repository

If the content repository disk is filled more than 50% (or any other value that is set in nifi.properties for nifi.content.repository.archive.max.usage.percentage), and if there is no data in the content repository archive, the following warning message can be found in the logs: "Unable to write flowfile content to content repository container default due to archive file size constraints; waiting for archive cleanup". This would block the processors and no more data is processed.

This appears to only happen if there is already data in the content repository on startup that needs to be archived, or if the following message is logged: “Found unknown file XYZ in the File System Repository; archiving file”.

Upstream JIRA
Action required
  • Upgrade (recommended)
    Upgrade to a version containing the fix.
    • CFM-2.2.5.2 for CDP 7.2.15.2
    • CFM-2.2.4.2 for CDP 7.2.14.3
    • CFM-2.2.3.4 for CDP 7.2.12.7
  • Workaround
    • Increase the value associated to nifi.content.repository.archive.max.usage.percentage (example: 75%).

      OR

    • Reduce disk space usage to get under the configured threshold.
Knowledge article
For the latest update on this issue see the corresponding Knowledge article: TSB 2022-580: NiFi Processors cannot write to content repository