CDE hardware requirements
Review the requirements needed to get started with the Cloudera Data Engineering (CDE) service on Red Hat OpenShift.
Requirements
- CDE assumes it has cluster-admin privileges on the OpenShift cluster.
- Openshift cluster should be configured with route admission policy set to namespaceOwnership: InterNamespaceAllowed. This allows Openshift cluster to run applications in multiple namespaces with the same domain name.
- CDE Service requirements:
Table 1. The following are the CDE Service requirements: Component vCPU Memory Block PV or NFS PV Number of pods Embedded DB -- -- 100 GB 1 Config Manager -- -- -- 2 Dex Downloads -- -- -- 1 Knox 250m 1 GB -- 1 Management API 1 2 GB -- 1 NGINX Ingress Controller 100m 90 MiB -- 1 FluentD Forwarder 1 1 GB -- 1 Grafana - - 10 GB 1 - CDE Virtual Cluster requirements: Overall storage 500 GB per Virtual Cluster and a storage
class named nfs exists backed by an NFS server.
Table 2. The following are the CDE Virtual Cluster requirements: Component vCPU Memory Block PV or NFS PV Number of pods Airflow API 1,250 m 2 GB, 512 MiB 100 GB 2 Airflow Scheduler 1 1 GB 100 GB 1 Airflow Web 1 2 GB 100 GB 1 Runtime API 1 2 GB 100 GB 1 Livy 3 12 GB 100 GB 1 Safari 4 6 GB -- 1 - Workloads: Depending upon the workload, you must configure resources. Minimal spark/airflow jobs need 1 executor with 1 CPU and 1GB Memory.