Known issues
Cloudbreak 2.9.2 includes the following known issues:
Note | |
---|---|
As of December 31, 2021, Cloudbreak reached end of
support. For more information, see Support lifecycle policy. Cloudera recommends that
you migrate your workloads to CDP Public Cloud. |
Known issues: Cloudbreak
Issue | Description | Workaround |
---|---|---|
BUG-117004 | When defining network security rules during cluster creation on
Azure, when ICMP protocol is used, cluster creation fails with an error
similar
to:
|
When defining network security rules during cluster creation on Azure, do not use the ICMP protocol. |
BUG-117005 | When defining network security rules during cluster creation on
Google Cloud via CLI, when ICMP protocol is specified and a port is
specified, cluster creation fails with an error similar
to: This
is because when the ICMP protocol is used, no ports should be specified.
The UI already enforces this automatically, but with CLI it is possible
to specify a port with the ICMP protocol. |
When defining network security rules during cluster creation on Google Cloud via CLI, if you would like to define a rule for using the ICMP protocol, do not specify any ports. |
BUG-110998 | When creating a cluster, the Cloud Storage page in the create cluster wizard includes an option to provide "Path to Ranger Audit Logs for Hive Property" when "Configure Storage Locations" is enabled. This option should only be available for data lakes and not for workload clusters. | Click on "Do not configure". |
BUG-99581 | The Event History in the Cloudbreak web UI displays the
following message: Manual recovery is needed for the following failed nodes: [] This message is displayed when Ambari agent doesn't send the heartbeat and Cloudbreak thinks that the host is unhealthy. However, if all services are green and healthy in Ambari web UI, then it is likely that the status displayed by Cloudbreak is incorrect. |
If all services are green and healthy in Ambari web UI, then syncing the cluster should fix the problem. |
BUG-110999 | The auto-import of HDP/HDF images on OpenStack does not work. This means, that in order to start creating HDP or HDF clusters on OpenStack, your OpenStack admin must import these images manually. | Your OpenStack admin must import these images manually by using the instructions in Import HDP and HDF images to OpenStack. |
BUG-112787 | When a cluster with the same name as specified in CLI JSON already
exists, CLI
returns:ERROR: status code: 403, message: Access is denied. |
To avoid this error, pass the cluster name as a parameter with
cb cluster create instead of including cluster name
in the CLI JOSN definition. |
Known issues: HDP
The known issues described here were discovered when testing Cloudbreak with HDP versions that are used by default in Cloudbreak. For general HDP known issues, refer to HDP release notes published at https://docs.hortonworks.com/.
There are no known issues related to HDP.
Known issues: HDF
The known issues described here were discovered when testing Cloudbreak with HDF versions that are used by default in Cloudbreak. For general HDF known issues, refer to HDF release notes published at https://docs.hortonworks.com/.
Issue | Description | Workaround |
---|---|---|
BUG-98865 | Blueprint configuration parameters are not applied when scaling an
HDF cluster. One example that affects all users is that after HDF cluster
upscale/downscale the nifi.web.proxy.host blueprint
parameter does not get updated to include the new nodes, and as a result
the NiFi UI is not reachable from these nodes. |
Configuration parameters set in the blueprint are not applied when
scaling an HDF cluster. One example that affects all NiFi users is that
after HDF cluster upscale the nifi.web.proxy.host
parameter does not get updated to include the new hosts, and as a result
the NiFi UI is not reachable from these
hosts.
|
Known issues: Data lake
Issue | Description | Workaround |
---|---|---|
BUG-109369 | Hive does not start on a HDP 2.6 data lake when Kerberos is enabled. |
|
BUG-116913, BUG-114150 | HiveServer2 does not start on an HDP 3.1 cluster attached to a data
lake. The following error is printed to Ambari
logs:
|
|