Deploying Kafka
You deploy a Kafka cluster by creating a Kafka resource and one or more KafkaNodePool resources in the Kubernetes cluster. The Kafka cluster can use either KRaft (recommended) or ZooKeeper (deprecated) for metadata management. After cluster deployment you can validate your cluster with the console producer and consumer tools shipped with Kafka.
- Kafka configuration that is common for the whole Kafka cluster (Kafka version, cluster name, and so on)
- Cruise Control configuration
- Entity Operator configuration
- ZooKeeper configuration (if ZooKeeper is used instead of KRaft)
A KafkaNodePool resource refers to a distinct group of Kafka nodes within a Kafka cluster. Using node pools enables you to specify different configurations for each node within the same Kafka cluster. Configuration options not specified in the node pool are inherited from the Kafka configuration.
You can deploy a Kafka cluster with one or more node pools. The number of node pools you create depends on how many groups of Kafka brokers you want to have that have differing configurations. The node pool configuration includes mandatory and optional settings. Configuration for replicas, roles, and storage is mandatory.
KRaft versus ZooKeeper
You can deploy Kafka in either KRaft or ZooKeeper mode. However, Cloudera recommends that you deploy clusters in KRaft mode. This is because ZooKeeper-based clusters are deprecated. Additionally, ZooKeeper will be removed in a future release.
KRaft offers enhanced reliability, scalability, and throughput over ZooKeeper. Metadata operations are more efficient as they are directly integrated. Additionally, when using KRaft, you are no longer required to maintain ZooKeeper, which reduces operational overhead.
Kafka node roles and deployment modes in KRaft
When you deploy a Kafka cluster in KRaft mode, you assign roles to each node in the Kafka cluster. Roles are assigned in the KafkaNodePool resource. There are two roles, broker and controller.
-
Broker ‐ These nodes manage Kafka records stored in topic partitions. Nodes with the broker role are your Kafka brokers.
-
Controller ‐ These nodes manage cluster metadata and the state of the cluster using a Raft-based consensus protocol. Controller nodes are the KRaft equivalent of ZooKeeper nodes.
A single Kafka node can have a single role or both roles. If you assign both roles to the node, it performs both broker and controller tasks. Depending on role assignments, your cluster will be running in one of the following modes.
-
KRaft mode ‐ In this mode, each Kafka node is either a broker or controller. Recommended for production clusters.
-
KRaft combined mode ‐ In this mode, some or all nodes in the cluster have both controller and broker roles assigned to them.
Combined mode is not recommended or supported for production environments. Use combined mode in development environments. Cloudera recommends that you always fully separate controller and broker nodes to avoid resource contention between roles.
Deploying a Kafka cluster in KRaft mode
You deploy a Kafka cluster in KRaft mode by deploying a Kafka resource and at least two KafkaNodePool resources. One KafkaNodepool describes your brokers, the other describes KRaft controllers. The Kafka resource must include the strimzi.io/kraft="enabled" annotation. Deploying Kafka in KRaft mode is the recommended mode for deployment.
-
Ensure that the Strimzi Cluster Operator is installed and running. See Installation.
-
Ensure that a namespace is available where you can deploy your cluster. If not, create one.
kubectl create namespace[***NAMESPACE***]
-
Ensure that the Secret containing credentials for the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted is available in the namespace where you plan on deploying your cluster. If the secret is not available, create it.
kubectl create secret docker-registry [***SECRET NAME***] \ --docker-server [***REGISTRY***] \ --docker-username [***USERNAME***] \ --docker-password [***PASSWORD***] \ --namespace [***NAMESPACE***]
-
[***SECRET NAME***] must be the same as the name of the Secret containing registry credentials that you created during Strimzi installation.
-
Replace [***REGISTRY***] with the server location of the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted. If your Kubernetes cluster has internet access, use
container.repository.cloudera.com
. Otherwise, enter the server location of your self-hosted registry. -
Replace [***USERNAME***] and [***PASSWORD***] with credentials that provide access to the registry. If you are using
container.repository.cloudera.com
, use your Cloudera credentials. Otherwise, enter credentials providing access to your self-hosted registry.
-
-
Scaling node pools that include KRaft controllers (controller roles) is not possible.
-
The following steps contain Kafka and KafkaNodePool resource examples. You can find additional examples on the Cloudera Archive.
Deploying a Kafka cluster in KRaft combined mode
You deploy a Kafka cluster in KRaft combined mode by deploying a Kafka resource and one or more KafkaNodePool resources. Typically you create two node pools, one describing nodes with both roles, and one that describes nodes that have the broker role only. Alternatively, you can create clusters where all nodes have both roles. In this case, a single node pool is sufficient. The Kafka resource must include the strimzi.io/kraft="enabled" annotation.
-
Ensure that the Strimzi Cluster Operator is installed and running. See Installation.
-
Ensure that a namespace is available where you can deploy your cluster. If not, create one.
kubectl create namespace[***NAMESPACE***]
-
Ensure that the Secret containing credentials for the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted is available in the namespace where you plan on deploying your cluster. If the secret is not available, create it.
kubectl create secret docker-registry [***SECRET NAME***] \ --docker-server [***REGISTRY***] \ --docker-username [***USERNAME***] \ --docker-password [***PASSWORD***] \ --namespace [***NAMESPACE***]
-
[***SECRET NAME***] must be the same as the name of the Secret containing registry credentials that you created during Strimzi installation.
-
Replace [***REGISTRY***] with the server location of the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted. If your Kubernetes cluster has internet access, use
container.repository.cloudera.com
. Otherwise, enter the server location of your self-hosted registry. -
Replace [***USERNAME***] and [***PASSWORD***] with credentials that provide access to the registry. If you are using
container.repository.cloudera.com
, use your Cloudera credentials. Otherwise, enter credentials providing access to your self-hosted registry.
-
-
Scaling node pools that include KRaft controllers (controller roles) is not possible.
Because of this limitation, you can only scale clusters running in combined mode if the cluster includes a node pool that has broker nodes only. The examples in these steps set up a broker-only node pool.
-
Ranger authorization does not work with combined mode.
-
The following steps contain Kafka and KafkaNodePool resource examples. You can find additional examples on the Cloudera Archive.
Deploying a Kafka cluster in ZooKeeper mode
You deploy a Kafka cluster in ZooKeeper mode by deploying a Kafka resource and at least a single KafkaNodePool resource. The Kafka resource must include ZooKeeper configuration.
-
Ensure that the Strimzi Cluster Operator is installed and running. See Installation.
-
Ensure that a namespace is available where you can deploy your cluster. If not, create one.
kubectl create namespace[***NAMESPACE***]
-
Ensure that the Secret containing credentials for the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted is available in the namespace where you plan on deploying your cluster. If the secret is not available, create it.
kubectl create secret docker-registry [***SECRET NAME***] \ --docker-server [***REGISTRY***] \ --docker-username [***USERNAME***] \ --docker-password [***PASSWORD***] \ --namespace [***NAMESPACE***]
-
[***SECRET NAME***] must be the same as the name of the Secret containing registry credentials that you created during Strimzi installation.
-
Replace [***REGISTRY***] with the server location of the Docker registry where Cloudera Streams Messaging - Kubernetes Operator artifacts are hosted. If your Kubernetes cluster has internet access, use
container.repository.cloudera.com
. Otherwise, enter the server location of your self-hosted registry. -
Replace [***USERNAME***] and [***PASSWORD***] with credentials that provide access to the registry. If you are using
container.repository.cloudera.com
, use your Cloudera credentials. Otherwise, enter credentials providing access to your self-hosted registry.
-
-
The following steps contain Kafka and KafkaNodePool resource examples. You can find additional examples on the Cloudera Archive.
Validating a Kafka cluster
After the Kafka broker pods are successfully started, you can use the Kafka console producer and consumer to validate the cluster. The following steps use the exact same docker images that were used to deploy the Kafka cluster by the Strimzi Cluster Operator. The images contain all the Kafka built-in tools and you can start a custom Kubernetes pod, starting the Kafka tools in the containers.