Known issues and limitations in Cloudera Data Engineering on CDP Private Cloud
This page lists the current known issues and limitations that you might run into while using the Cloudera Data Engineering (CDE) service.
- DEX-14676: Deep Analysis is not working in CDE PvC under analysis tab
- If you are using Spark version 2.x for running your jobs, then the Run Deep Analysis feature present under the Analysis tab is not supported on Cloudera Data Engineering Private Cloud.
- DEX-8540: Job Analysis tab is not working
- When you access the Analysis tab fails to load data for Spark 2. tab through the Cloudera Data Engineering UI, the
- DEX-8682: CDE PvC 1.5.0 : CDP upgrade to 1.5.0 with OCP upgrade (4.8 to 4.10) Jobs UI is not opening
- Upgrading the OCP version from 4.8 to 4.10 while CDE service is enabled, causes the Jobs UI to not open. This is due to OCP 4.10 upgrading to the Kubernetes version 1.23 which removes the old ingress APIs used.
- DOCS-17844: Logs are lost if the log lines are longer than 50000 characters in fluentd
-
This issue occurs when the Buffer_Chunk_Size parameter for the fluent-bit is set to a value that is lesser than the size of the log line.
- DEX-8614: Sometimes Spark job is not getting killed even though its parent Airflow job gets killed
- Sometimes if an issue is encountered while sending the request to kill a spark batch to the Livy API and the error is logged but not propagated properly to the Airflow job. In such cases, the underlying spark job might still be running, though the airflow job considers that the job is killed successfully.
- DEX-9237: Job fails with the “Permission Denied” error after updating the virtual cluster resource quota
- Whenever the virtual cluster resource quota is updated, the newly launched jobs on the virtual cluster fail with the Permission Denied error. This error can be seen in various stages of the job life cycle, in submitters, drivers or executors, and Airflow workers.
- DEX-8601: ECS 1.4.x to 1.5.0 Upgrade: jobs fail after upgrade
- Upgrading the ECS version while CDE service is enabled, causes the jobs launched in the old CDE virtual cluster fail. This is due to ECS upgrading to the kubernetes version 1.23 which removes the old ingress APIs used.
- DEX-8600: ECS 1.4.x to 1.5.0 Upgrade: Virtual cluster creation and deletion fails
- Upgrading the ECS version while CDE service is enabled, causes the old CDE service and virtual cluster creation and deletion to fail. This is due to ECS upgrading to the kubernetes version 1.23 which removes the old ingress APIs used.
- DEX-8226: Grafana Charts of new virtual clusters will not be accessible on upgraded clusters if virtual clusters are created on existing CDE service.
- If you upgrade the cluster from 1.3.4 to 1.4.x and create a new virtual clusters on the existing CDE Service, Grafana Charts will not be displayed. This is due to broken APIs.
- DEX-7000: Parallel Airflow tasks triggered at exactly same time by the user throws the 401:Unauthorized error.
- Error
401:Unauthorized
causes airflow jobs to fail intermittently, when parallel Airflow tasks usingCDEJobRunOperator
are triggered at the exact same time in an Airflow DAG. - DEX-7001: When Airflow jobs are run, the privileges of the user who created the job is applied and not the user who submitted the job.
- Irrespective of who submits the Airflow job, the Airflow job is run with the user privileges who created the job. This causes issues when the job submitter has lesser privileges than the job owner who has higher privileges.
- Changing LDAP configuration after installing CDE breaks authentication
- If you change the LDAP configuration after installing CDE, as described in Configuring LDAP authentication for CDP Private Cloud, authentication no longer works.
- HDFS is the default filesystem for all resource mounts
- For any jobs that use local filesystem paths as
arguments to a Spark job, explicitly specify
file://
as the scheme. For example, if your job uses a mounted resource calledtest-resource.txt
, in the job definition, you would typically refer to it as/app/mount/test-resource.txt
. In CDP Private Cloud, this should be specified asfile:///app/mount/test-resource.txt
. - Apache Ozone is supported only for log files
- Apache Ozone is supported only for log files. It is not supported for job configurations, resources, and so on.
- Scheduling jobs with URL references does not work
- Scheduling a job that specifies a URL reference does not work.
Limitations
- Access key-based authentication will not be enabled in upgraded clusters prior to CDP PVC 1.3.4 release.
- After you upgrade to PVC 1.3.4 version from earlier versions, you must create the CDE Base service and Virtual Cluster again to use the new Access Key feature. Otherwise, the Access Key feature will not be supported in the CDE Base service created prior to the 1.3.4 upgrade.