Cloudera Manager 7.11.3 Cumulative hotfix 10
Know more about the Cloudera Manager 7.11.3 cumulative hotfixes 10.
This cumulative hotfix was released on November 12, 2024.
- Enhancements to the Observability page
- The following changes have been made to the Observability
page::
- Added role-specific metrics to the Status and Charts Library tabs for component servers such as Pipelines, ADB, and SDX.
- Added relevant metrics across all Observability component servers to the Status and Charts Library tabs for the Observability page.
- ENGESC-30503, OPSAPS-74868: Cloudera Manager limited support for custom external repository requiring basic authentication
- Current Cloudera Manager does not support custom external
repository with basic authentication (the Cloudera Manager Wizard supports either HTTP
(non-secured) repositories or usage of Cloudera
https://archive.cloudera.com
only). In case customers want to use a custom external repository with basic authentication, they might get errors. - OPSAPS-60726: Newly saved parcel URLs are not showing up in the parcels page in the Cloudera Manager HA cluster.
- To safely manage parcels in a Cloudera Manager HA
environment, follow these steps:
- Shutdown the Passive Cloudera Manager Server.
- Add and manage the parcel as usual, as described in Install Parcels.
- Restart the Passive Cloudera Manager server after parcel operations are complete.
- OPSAPS-73211: Cloudera Manager 7.11.3 does not clean up Python Path impacting Hue to start
-
When you upgrade from Cloudera Manager 7.7.1 or lower versions to Cloudera Manager 7.11.3 or higher versions with CDP Private Cloud Base 7.1.7.x Hue does not start because Cloudera Manager forces Hue to start with Python 3.8, and Hue needs Python 2.7.
The reason for this issue is because Cloudera Manager does not clean up the Python Path at any time, so when Hue tries to start the Python Path points to 3.8, which is not supported in CDP Private Cloud Base 7.1.7.x version by Hue.
- OPSAPS-73011: Wrong parameter in the /etc/default/cloudera-scm-server file
- In case the Cloudera Manager needs to be installed in
High Availability (2 nodes or more as explained here), the parameter
CMF_SERVER_ARGS
in the /etc/default/cloudera-scm-server file is missing the word "export
" before it (on the file there is onlyCMF_SERVER_ARGS=
and notexport CMF_SERVER_ARGS=
), so the parameter cannot be utilized correctly. - OPSAPS-72984: Alerts due to change in hostname fetching functionality in jdk 8 and jdk 11
-
Upgrading JAVA from JDK 8 to JDK 11 creates the following alert in CMS:
Bad : CMSERVER:pit666.slayer.mayank: Reaching Cloudera Manager Server failed
This happens due to a functionality change in JDK 11 on hostname fetching.[root@pit666.slayer ~]# /us/lib/jvm/java-1.8.0/bin/java GetHostName Hostname: pit666.slayer.mayank [root@pit666.slayer ~]# /usr/lib/jvm/java-11/bin/java GetHostName Hostname: pit666.slayer
You can notice that the "hostname" is set to a short name instead of
FQDN
. - OPSAPS-73655: Cloud replication fails after the delegation token is issued
- HDFS and Hive external table replication policies from an
on-premises cluster to cloud fail when the following conditions are true:
- You choose the option during the replication policy creation process.
- Incremental replication is in progress, that is the source paths of the replication are snapshottable directories and the bootstrap replication run is complete.
- OPSAPS-65377: Cloudera Manager - Host Inspector not finding Psycopg2 on Ubuntu 20 or Redhat 8.x when Psycopg2 version 2.9.3 is installed.
-
Host Inspector fails with Psycopg2 version error while upgrading to Cloudera Manager 7.11.3 or Cloudera Manager 7.11.3 CHF-x versions. When you run the Host Inspector, you get an error Not finding Psycopg2, even though it is installed on all hosts.
- OPSAPS-68340: Zeppelin paragraph execution fails with the User not allowed to impersonate error.
-
Starting from Cloudera Manager 7.11.3, Cloudera Manager auto-configures the
livy_admin_users
configuration when Livy is run for the first time. If you add Zeppelin or Knox services later to the existing cluster and do not manually update the service user, the User not allowed to impersonate error is displayed. - OPSAPS-69847:Replication policies might fail if source and target use different Kerberos encryption types
-
Replication policies might fail if the source and target Cloudera Manager instances use different encryption types in Kerberos because of different Java versions. For example, the Java 11 and higher versions might use the aes256-cts encryption type, and the versions lower than Java 11 might use the rc4-hmac encryption type.
- OPSAPS-69342: Access issues identified in MariaDB 10.6 were causing discrepancies in High Availability (HA) mode
-
MariaDB 10.6, by default, includes the property
require_secure_transport=ON
in the configuration file (/etc/my.cnf), which is absent in MariaDB 10.4. This setting prohibits non-TLS connections, leading to access issues. This problem is observed in High Availability (HA) mode, where certain operations may not be using the same connection. - OPSAPS-70771: Running Ozone replication policy does not show performance reports
- During an Ozone replication policy run, the A
server error has occurred. See Cloudera Manager server log for details error
message appears when you click:
- Replication Policies page. or on the
- Download CSV on the Replication History page to download any report.
- OPSAPS-70713: Error appears when running Atlas replication policy if source or target clusters use Dell EMC Isilon storage
- You cannot create an Atlas replication policy between clusters if one or both the clusters use Dell EMC Isilon storage.
- CDPD-53185: Clear REPL_TXN_MAP table on target cluster when deleting a Hive ACID replication policy
- The entry in REPL_TXN_MAP table on the target cluster is
retained when the following conditions are true:
- A Hive ACID replication policy is replicating a transaction that requires multiple replication cycles to complete.
- The replication policy and databases used in it get deleted on the source and target cluster even before the transaction is completely replicated.
In this scenario, if you create a database using the same name as the deleted database on the source cluster, and then use the same name for the new Hive ACID replication policy to replicate the database, the replicated database on the target cluster is tagged as ‘database incompatible’. This happens after the housekeeper thread process (that runs every 11 days for an entry) deletes the retained entry.
- OPSAPS-71592: Replication Manager does not read the default value of “ozone_replication_core_site_safety_valve” during Ozone replication policy run
- During the Ozone replication policy run, Replication
Manager does not read the value in the
ozone_replication_core_site_safety_valve
advanced configuration snippet if it is configured with the default value. - OPSAPS-72468: Subsequent Ozone OBS-to-OBS replication policy do not skip replicated files during replication
- The first Ozone replication policy run is a bootstrap run. Sometimes, the subsequent runs might also be bootstrap jobs if the incremental replication fails and the job runs fall back to bootstrap replication. In this scenario, the bootstrap replication jobs might replicate the files that were already replicated because modification time is different for a file on the source and the target cluster.
- OPSAPS-72509, CDPD-32440: Hive metadata transfer to GCS fails with ClassNotFoundException
- Hive replication policies from an on-premises cluster to
cloud fails during the “Transfer Metadata Files” step if the following conditions are
true:
- the target is a GCS Data Lake
- the source Cloudera Manager version is 7.11.3 CHF7, 7.11.3 CHF8, 7.11.3 CHF9, 7.11.3 CHF9.1, 7.11.3 CHF10, or 7.11.3 CHF11
fs.gs.delegation.token.binding
property is already defined in the configuration and cannot be unset to disable the delegation tokens in the cloud connector service. - OPSAPS-74398: Ozone and HDFS replication policies might fail when you use different destination proxy user and source proxy user
- HDFS on-premises to on-premises replication fails when
the following conditions are true:
- You configure different Run As Username and Run on Peer as Username during the replication policy creation process.
- The user configured in Run As Username does not have the permission to access the source path on the source HDFS.
- OPSAPS-71642:
GflagConfigFileGenerator
is removing the=
sign in the Gflag configuration file when the configuration value passed is empty in the advanced safety valve -
If the user adds
file_metadata_reload_properties
configuration in the advanced safety valve with=
sign and empty value, then theGflagConfigFileGenerator
is removing the=
sign in the Gflag configuration file when the configuration value passed is empty in the advanced safety valve.This issue is fixed now.
- OPSAPS-70704: Kerberos connectivity check does not work as expected with JDK17 when you add Cloudera Manager peers
- When you add a source Cloudera Manager that supports JDK17, the Kerberos connectivity check fails and the Error while reading /etc/krb5.conf on <hostname ; for all hosts>... error message appears. This issue is fixed now.
- OPSAPS-71666: Replication Manager uses the required property values in the “ozone_replication_core_site_safety_valve ” in the source Cloudera Manager during Ozone replication policy run
- During an Ozone replication policy run, Replication
Manager obtains the required properties and its values from the
ozone_replication_core_site_safety_valve
. It then adds the new properties and its values and overrides the value for existing properties in the core-site.xml file. Replication Manager uses this file during the Ozone replication policy run. - OPSAPS-71647: Ozone replication fails for incompatible source and target Cloudera Manager versions during the payload serialization operation
- Replication Manager now recognizes and annotates the required fields during the payload serialization operation. For the list of unsupported Cloudera Manager versions that do not have this fix, see Preparing clusters to replicate Ozone data.
- OPSAPS-71615: Service monitor crashes due to an out-of-memory (OOM) error.
- This issue is fixed now.
- OPSAPS-71258: Kafka, SRM, and SMM cannot process messages compressed with Zstd or Snappy
- Kafka, SRM, and SMM cannot process messages compressed with Zstd or Snappy if the /tmp directory is mounted with noexec flag. This issue is fixed now.
- OPSAPS-70848: Hive external table replication policies succeed when the source cluster uses Dell EMC Isilon storage
- During the Hive external table replication policy run,
the replication policy failed at the
Hive Replication Export
step. This issue is resolved. - OPSAPS-70822: Save the Hive external table replication policy on the ‘Edit Hive External Table Replication Policy’ window
- Replication Manager saves the changes as expected when you click Save Policy after you edit a Hive replication policy. To edit a replication policy, you click for the replication policy on the Replication Policies page.
- OPSAPS-69782: Exception appears if the peer Cloudera Manager's API version is higher than the local cluster's API version
- HBase replication using HBase replication policies in CDP
Public Cloud Replication Manager between two Data Hubs/COD clusters succeed as expected
when all the following conditions are true:
- The destination Data Hub/COD cluster’s Cloudera Manager version is 7.9.0-h7 through 7.9.0-h9 or 7.11.0-h2 through 7.11.0-h4, or 7.12.0.0.
- The source Data Hub/COD cluster's Cloudera Manager major version is higher than the destination cluster's Cloudera Manager major version.
- The Initial Snapshot option is chosen during the HBase replication policy creation process and/or the source cluster is already participating in another HBase replication setup as a source or destination with a third cluster.
- OPSAPS-66459: Enable concurrent Hive external table replication policies with the same cloud root
- When the
HIVE_ALLOW_CONCURRENT_REPLICATION_WITH_SAME_CLOUD_ROOT_PATH
feature flag is enabled, Replication Manager can run two or more Hive external table replication policies with the same cloud root path concurrently.For example, if two Hive external table replication policies have s3a://bucket/hive/data as the cloud root path and the feature flag is enabled, Replication manager runs these policies concurrently.
By default, this feature flag is disabled. To enable the feature flag, contact your Cloudera account team.
- Fixed Common Vulnerabilities and Exposures
- For information about Common Vulnerabilities and Exposures (CVE) that are fixed in Cloudera Manager 7.11.3 cumulative hotfix 10, see Fixed Common Vulnerabilities and Exposures in Cloudera Manager 7.11.3 cumulative hotfixes.
The repositories for Cloudera Manager 7.11.3-CHF 10 are listed in the following table:
Repository Type | Repository Location |
---|---|
RHEL 9 Compatible | Repository: Repository
File:
|
RHEL 8 Compatible | Repository: Repository
File:
|
RHEL 7 Compatible | Repository: Repository
File:
|
SLES 15 | Repository: Repository
File:
|
SLES 12 | Repository: Repository
File:
|
Ubuntu 20 | Repository: Repository
File:
|
Ubuntu 22 | Repository: Repository
File:
|