Known Issues in MapReduce and YARN
Learn about the known issues in Mapreduce and YARN, the impact or changes to the functionality, and the workaround.
Known Issues
- COMPX-7586: Max Parallel Apps cannot be changed for root queue
-
The Queue Manager UI may show incorrect value for the Maximum Parallel Applications property for the root queue. The value can be changed with the QM UI, but the UI will always show 2147483647.
- COMPX-5817: Queue Manager UI will not be able to present a view of pre-upgrade queue structure. CM Store is not supported and therefore Yarn will not have any of the pre-upgrade queue structure preserved.
- When a Data Hub cluster is deleted, all saved configurations are also deleted. All YARN configurations are saved in CM Store and this is yet to be supported in Data Hub and Cloudera Manager. Hence, the YARN queue structure also will be lost when a Data Hub cluster is deleted or upgraded or restored.
- COMPX-5244: Root queue should not be enabled for auto-queue creation
- After dynamic auto child creation is enabled for a queue using the YARN Queue Manager UI, you cannot disable it using the YARN Queue Manager UI. That can cause problem when you want to switch between resource allocation modes, for example from weight mode to relative mode. The YARN Queue Manager UI does not let you to switch resource allocation mode if there is at least one dynamic auto child creation enabled parent queue in your queue hierarchy.
- COMPX-5589: Unable to add new queue to leaf queue with partition capacity in Weight/Absolute mode
- Scenario
- User creates one or more partitions.
- Assigns a partition to a parent with children
- Switches to the partition to distribute the capacities
- Creates a new child queue under one of the leaf queues but the following error is displayed:
Error : 2021-03-05 17:21:26,734 ERROR com.cloudera.cpx.server.api.repositories.SchedulerRepository: Validation failed for Add queue operation. Error message: CapacityScheduler configuration validation failed:java.io.IOException: Failed to re-init queues : Parent queue 'root.test2' have children queue used mixed of weight mode, percentage and absolute mode, it is not allowed, please double check, details: {Queue=root.test2.test2childNew, label= uses weight mode}. {Queue=root.test2.test2childNew, label=partition uses percentage mode}
- COMPX-5264: Unable to switch to Weight mode on creating a managed parent queue in Relative mode
- In the current implemention, if there is an existing managed queue in Relative mode, then conversion to Weight mode is not be allowed.
- COMPX-5549: Queue Manager UI sets maximum-capacity to null when you switch mode with multiple partitions
- If you associate a partition with one or more queues and then switch the allocation mode before assigning capacities to the queues, an Operation Failed error is displayed as the max-capacity is set to null.
- COMPX-4992: Unable to switch to absolute mode after deleting a partition using YARN Queue Manager
- If you delete a partition (node label) which has been associated with queues and those queues have capacities configured for that partition (node label), the CS.xml still contains the partition (node label) information. Hence, you cannot switch to absolute mode after deleting the partition (node label).
- COMPX-3181: Application logs does not work for AZURE and AWS cluster
- Yarn Application Log Aggregation will fail for any YARN job (MR, Tez, Spark, etc) which do not use cloud storage, or use a cloud storage location other than the one configured for YARN logs (yarn.nodemanager.remote-app-log-dir).
- COMPX-1445: Queue Manager operations are failing when Queue Manager is installed separately from YARN
- If Queue Manager is not selected during YARN installation, Queue Manager operation are failing. Queue Manager says 0 queues are configured and several failures are present. That is because ZooKeeper configuration store is not enabled.
- COMPX-1451: Queue Manager does not support multiple ResourceManagers
- When YARN High Availability is enabled there are multiple ResourceManagers. Queue Manager receives multiple ResourceManager URLs for a High Availability cluster. It picks the active ResourceManager URL only when Queue Manager page is loaded. Queue Manager cannot handle it gracefully when the currently active ResourceManager goes down while the user is still using the Queue Manager UI.
- COMPX-3329: Autorestart is not enabled for Queue Manager in Data Hub
- In a Data Hub cluster, Queue Manager is installed with autorestart disabled. Hence, if Queue Manager goes down, it will not restart automatically.
- Third party applications do not launch if MapReduce framework path is not included in the client configuration
- MapReduce application framework is loaded from HDFS instead of being present on the
NodeManagers. By default the
mapreduce.application.framework.path
property is set to the appropriate value, but third party applications with their own configurations will not launch. - OPSAPS-57067: Yarn Service in Cloudera Manager reports stale configuration yarn.cluster.scaling.recommendation.enable.
- This issue does not affect the functionality. Restarting Yarn service will fix this issue.
- JobHistory URL mismatch after server relocation
- After moving the JobHistory Server to a new host, the URLs listed for the JobHistory Server on the ResourceManager web UI still point to the old JobHistory Server. This affects existing jobs only. New jobs started after the move are not affected.
- CDH-49165: History link in ResourceManager web UI broken for killed Spark applications
- When a Spark application is killed, the history link in the ResourceManager web UI does not work.
- CDH-6808: Routable IP address required by ResourceManager
- ResourceManager requires routable
host:port
addresses foryarn.resourcemanager.scheduler.address
, and does not support using the wildcard 0.0.0.0 address. - OPSAPS-52066: Stacks under Logs Directory for Hadoop daemons are not accessible from Knox Gateway.
- Stacks under the Logs directory for Hadoop daemons, such as NameNode, DataNode, ResourceManager, NodeManager, and JobHistoryServer are not accessible from Knox Gateway.
- CDPD-2936: Application logs are not accessible in WebUI2 or Cloudera Manager
- Running Containers Logs from NodeManager local directory cannot be accessed either in Cloudera Manager or in WebUI2 due to log aggregation.
- OPSAPS-50291: Environment variables
HADOOP_HOME
,PATH
,LANG
, andTZ
are not getting whitelisted - It is possible to include the environment variables
HADOOP_HOME
,PATH
,LANG
, andTZ
in the allowlist, but the container launch environments do not have these variables set up automatically. - YARN cannot start if Kerberos principal name is changed
- If the Kerberos principal name is changed in Cloudera Manager after launch, YARN will not be able to start. In such case the keytabs can be correctly generated but YARN cannot access ZooKeeper with the new Kerberos principal name and old ACLs.
- COMPX-8687: Missing access check for getAppAttemps
- When the Job ACL feature is enabled using Cloudera Manager (
mapreduce.cluster.acls.enabled
property is not generated to all configuration files, including theyarn-site.xml
configuration file. As a result the ResourceManager process will use the default value of this property. The default property ofmapreduce.cluster.acls.enabled
isfalse
.
property), the
Unsupported Features
-
The following YARN features are currently not supported in Cloudera Data Platform:
- Application Timeline Server (ATSv2 and ATSv1)
- Container Resizing
- Distributed or Centralized Allocation of Opportunistic Containers
- Distributed Scheduling
- Docker on YARN (DockerContainerExecutor) on Data Hub clusters
- Fair Scheduler
- GPU support for Docker
- Hadoop Pipes
- Moving jobs between queues
- Native Services
- Pluggable Scheduler Configuration
- Queue Priority Support
- Reservation REST APIs
- Resource Estimator Service
- Resource Profiles
- (non-Zookeeper) ResourceManager State Store
- Rolling Log Aggregation
- Shared Cache
- YARN Federation