Configure the processor for your data source

You can set up a data flow to move data from many locations into Apache HBase. This example assumes that you are configuring ConsumeKafkaRecord_2_0. If you are moving data from a location other than Kafka, review Getting Started with Apache NiFi for information about how to build a data flow, and about other data consumption processor options.

  • Ensure that you have the Ranger policies required to access the Kafka consumer group.
  • This use case demonstrates a data flow using data in Kafka, and moving it to HBase. To review how to ingest data to Kafka, see Ingesting Data to Apache Kafka.
  1. Launch the Configure Processor window, by right-clicking the processor and selecting Configure.
  2. Click the Properties tab.
  3. Configure ConsumeKafkaRecord_2_0 with the required values.

    The following table includes a description and example values for the properties required to configure an ingest data flow. For a complete list of ConsumeKafkaRecord_2_0, see the processor documentation.

    Property Description Value for ingest to HBase data flow
    Kafka Brokers Provide a comma-separated list of known Kafka Brokers in the format <host>:<port>.,,
    Topic Name(s) Enter the name of the Kafka topic from which you want to get data. customer
    Topic Name Format Specify whether the topics are a comma separated list of topic names, or a single regular expression.


    Record Reader Specifies the record reader to use for incoming Flow Files. This can be a range of data formats.


    Record Writer

    Specifies the format you want to use in order to serialize data before for sending it to the data target.

    Note: Ensure that the source record writer and the target record reader are using the same Schema.


    Honor Transactions Specifies whether or not NiFi should honor transactional guarantees when communicating with Kafka.


    Security Protocol Specifies the protocol used to communicate with brokers. This value corresponds to Kafka's 'security.protocol' property.


    SASL Mechanism The SASL mechanism to use for authentication. Corresponds to Kafka's 'sasl.mechanism' property.


    Username Specify the username when the SASL Mechanism is PLAIN or SCRAM-SHA-256.


    Password Specify the password for the given username when the SASL Mechanism is PLAIN or SCRAM-SHA-256.


    SSL Context Service Specifies the SSL Context Service to use for communicating with Kafka.


    Group ID A Group ID is used to identify consumers that are within the same consumer group. Corresponds to Kafka's '' property.


Configure the processor for your data target.