Using ReadyFlows

The ReadyFlow Gallery is where you can find out-of-box flow definitions. To use a ReadyFlow, add it to the Catalog and then use it to create a Flow Deployment.

Kafka to S3 Avro

Kafka to S3 Avro ReadyFlow Summary

This ReadyFlow consumes JSON, CSV or Avro data from a source Kafka topic and merges the events into Avro files before writing the data to S3. The flow writes out a file every time its size has either reached 100MB or five minutes have passed.

Ingesting Data using the Kafka to S3 Avro ReadyFlow

You can use the Kafka to S3 Avro ReadyFlow to move your data into an AWS bucket.

Kafka Filter to Kafka

Kafka Filter to Kafka ReadyFlow Summary

This ReadyFlow consumes JSON, CSV, or Avro data from a source Kafka topic and parses the schema by looking up the schema name in the CDP Schema Registry. You can filter events by specifying a SQL query in the Filter Rule parameter.

Ingesting Data using the Kafka Filter to Kafka ReadyFlow

You can use the Kafka Filter to Kafka ReadyFlow to move your data from a Kafka topic to another Kafka topic while applying a schema to the data in Cloudera DataFlow (CDF).

Kafka to Kudu

Kafka to Kudu ReadyFlow Summary

This ReadyFlow consumes JSON, CSV or Avro data from a source Kafka topic, parses the schema by looking up the schema name in the CDP Schema Registry and ingests it into a Kudu table.

Ingesting Data using the Kafka to Kudu ReadyFlow

You can use the Kafka to Kudu ReadyFlow to move your data from a Kafka topic into Apache Kudu in a CDP Public Cloud Real-time Data Mart cluster.