Configure the processor for your data source

Learn how you can configure the ConsumeKafkaRecord_2_0 data source processor for your ADLS ingest data flow. You can set up a data flow to move data to Azure Data Lake Storage from many different locations. This example assumes that you are streaming data from Kafka and shows you the configuration for the relevant data source processor.

  1. Launch the Configure Processor window, by right-clicking the ConsumeKafkaRecord_2_0 processor and selecting Configure.

    This gives you a configuration dialog with the following tabs: Settings, Scheduling, Properties, Comments.

  2. Configure the processor according to the behavior you expect in your data flow.
  3. When you have finished configuring the options you need, save the changes by clicking the Apply button.
    Make sure that you set all required properties, as you cannot start the processor until all mandatory properties have been configured.

In this example data flow, the data source is Kafka. You can create the modified Kafka broker URLs using the broker hostnames and adding port :9093 to the end of each FQDN. You can find the hostnames on the Streams Messaging cluster overview page when selecting the Hardware tab.

The following table includes a description and example values for the properties required to configure the example ingest data flow. For a complete list of ConsumeKafkaRecord_2_0 properties, see the Apache Nifi Documentation.

Table 1. ConsumeKafkaRecord_2_0 processor properties
Property Description Example value for ingest data flow

Kafka Brokers

Provide a comma-separated list of known Kafka Brokers.

In the format: <host>:<port>,,

Topic Name(s)

Provide the name of the Kafka Topic(s) to pull from.


Record Reader

Specify the Record Reader to use for incoming FlowFiles.


Record Writer

Specify the Record Writer to use in order to serialize the data before sending to Kafka.


Security Protocol

Specify the protocol used to communicate with Kafka brokers.


SASL Mechanism

Specify the SASL mechanism to use for authentication.



Use your CDP workload username to set this Authentication property.



Use your CDP workload password to set this Authentication property.


SSL Context Service

Specify the SSL Context Service to use for communicating with Kafka.

Use the pre-configured SSLContextProvider.

Default NiFi SSL Context Service

Group ID

Provide the consumer group ID to identify consumers that are within the same consumer group.


Configure the processor for merging your records.