Known Issues in Cloudera Data Explorer (Hue)

Known issues and technical limitations for Cloudera Data Explorer (Hue) are addressed in Cloudera Runtime 7.3.2, its service packs, and cumulative hotfixes.

Known issues identified in Cloudera Runtime 7.3.2

CDPD-98754: Data Explorer cannot access HDFS in Data Discovery Template
7.3.2
In the Data Discovery Template, Data Explorer cannot access HDFS. This issue occurs because Cloudera Manager assigns a standby HDFS NameNode to the webhdfs_url property instead of the active NameNode, which prevents the file browser from displaying HDFS data.
  1. Log in to Cloudera Manager as an Administrator.
  2. Go to Clusters > HDFS to identify which NameNode is currently active.
  3. Go to Clusters > Data Explorer (Hue) > Configuration.

  4. Change the value of the webhdfs_url property from the standby NameNode to the active NameNode.
  5. Click Save Changes.
  6. Restart the Data Explorer service.
CDPD-95911: File operations fail for files with special characters in GS RAZ environments
7.3.2
In GS RAZ environments, you cannot rename, move, or copy files that contain special characters such as the percent sign (%). When you attempt these operations, the Failed to receive a valid response from RAZ error message is displayed. Additionally, operations on files containing the hash character (#) do not result in an error, but the action is not performed.
None.
CDPD-94944: File and directory names contain special characters after a rename operation
7.3.2
In Data Explorer, file and directory names that contain special characters, such as the ampersand (&) and the apostrophe ('), are incorrectly transformed into HTML entities after a rename operation. For example, the ampersand is replaced by & and the apostrophe is replaced by '. This issue occurs during rename operations for files with names containing characters such as ~@$&()*!+'=;.txt.
None.
CDPD-95666: Endless progress bar for S3 folders without read permissions on Azure
7.3.2
In Azure environments, a persistent loading bar is displayed when you select a folder or file for which you do not have read permissions. This issue occurs across the file system, including S3 and HDFS. Currently, the user interface does not display an alert message to indicate that permissions are missing. Instead, the UI remains in a loading state.
None.
CDPD-93181: Incorrect Data Explorer server status reporting
Data Explorer server status is incorrectly reported as Green (Healthy) in Cloudera Manager even if the process fails to start. This occurs when the mysqlclient dependency is missing, causing database migrations to fail. The Data Explorer server appears healthy because the loglistener.py sidecar process starts first and maintains a Process ID (PID). Cloudera Manager monitors this PID and reports a successful start, even though the main Data Explorer process has terminated due to the missing dependency.
  1. Stop the Hue service in Cloudera Manager.
  2. Log in to the affected Hue host and manually terminate the loglistener.py sidecar process.
  3. Remove any stale PID files located in the /tmp/ directory, for example, /tmp/hue_*.pid.
  4. Ensure that the mysqlclient dependency is installed on the host before attempting to restart the service.

Known issues identified Before Cloudera Runtime 7.3.2

Known issues identified before Cloudera Runtime 7.3.2 include only unresolved issues from previous releases that continue to affect the Cloudera Runtime 7.3.2 base release.

CDPD-95086: Data Explorer UI fails to load due to a connection timeout error
7.3.2, 7.3.1.706
In a Cloudera Data Hub high availability (HA) environment, the Data Explorer UI can fail to load after the master0 HDFS node is deleted. Data Explorer continues to attempt connections to the deleted node because Data Explorer does not support HDFS HA. As a result, the Data Explorer file browser points to an invalid HDFS endpoint, which causes connection timeouts and renders the UI inaccessible.
You can restore access to the Data Explorer UI by updating the configured HttpFS node to an active HDFS node in Cloudera Manager and then restarting the Data Explorer service.
CDPD-92946: Data Explorer debug logging cannot be disabled
7.3.2, 7.3.1.600, 7.3.1.706
Data Explorer debug logging is always enabled and cannot be turned off.
None.
CDPD-88964: Data Explorer logs missing in Data Explorer UI
7.3.2, 7.3.1.500 through 7.3.1.706
The Data Explorer UI might display only a few lines of logs instead of the complete Data Explorer logs. This can occur due to leftover Gunicorn processes that interfere with the proper logging and display of logs within the Data Explorer interface.
  1. Stop the Data Explorer service.
  2. Terminate any remaining Gunicorn processes to clear hung or orphan processes that might be causing the issue. Run the following command (use sudo if not running as root):
    # pids=$(ps -efwww | grep rungunicornserver | grep -v "grep" | grep "rungunicornserver"| awk '{ print $2 }') && for i in $pids; do kill -9 $i ; done
  3. Restart the Data Explorer service.
CDPD-90510: Defunct Data Explorer Gunicorn worker processes accumulate
7.3.2, 7.3.1.500 through 7.3.1.706
Data Explorer on Ubuntu 22 using Oracle Database can accumulate defunct rungunicornserver worker processes due to incomplete process termination and stale database connections. This can lead to a cluttered process table, which does not critically impact service functionality.
Periodically clean defunct worker processes by running the following command:
pgrep -f 'hue rungunicornserver' | xargs -r kill -9
CDPD-58978: Batch query execution using Data Explorer fails with Kerberos error
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs
When you run Impala queries in a batch mode, you encounter failures with a Kerberos error even if the keytab is configured correctly. This is because submitting Impala, Sqoop, Pig, or pyspark queries in a batch mode launches a shell script Oozie job from Data Explorer and this is not supported on a secure cluster.
None. Submit the queries individually.
CDPD-54376: Clicking the home button on the File Browser page redirects to HDFS user directory
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs
When you are previewing a file on any supported filesystem, such as S3 or ABFS, and you click the Home button, you are redirected to the HDFS user home directory instead of the user home directory on the specific filesystem.
None.
CDPD-43293: Unable to import Impala table using Importer
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs, 7.2.16 and its SPs
Creating Impala tables using the Data Explorer Importer might fail.
  • If you have both Hive and Impala services installed on your cluster, then you can import the table using by selecting the Hive dialect from Tables > Sources.
  • If only Impala service is installed on your cluster, go to Cloudera Manager > Clusters > Hue > Configurations and add the following line in the Hue Service Advanced Configuration Snippet (Safety Valve) for hue_safety_valve.ini field:
    [beeswax]
    max_number_of_sessions=1
CDPD-64541, CDPD-63617: Creating managed tables using Data Explorer Importer fails on RAZ-enabled GCP environments
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs
On Google Cloud Platform (GCP) environments, creating managed tables using either Hive or Impala dialects fails. This process results in the creation of orphaned temporary (tmp) tables. The failure occurs because Hive and Impala cannot load data from the inpath in Google Storage when requested outside of the Hue interface.
None.
CDPD-56888: Renaming a folder with special characters results in a duplicate folder with a new name on AWS S3.
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs
On AWS S3, if you try to rename a folder with special characters in its name, a new folder is created as a copy of the original folder with its contents. Also, you might not be able to delete the folder containing special characters.
.
  1. SSH into your Cloudera environment host.
  2. Rename or delete a directory having special characters.
    • To delete a directory within your S3 bucket, run the following command:
      hdfs dfs -rm -r [***COMPLETE-PATH-TO-S3-BUCKET***]/[***DIRECTORY-NAME***]
    • To rename a folder, create a new directory and run the following command to move files from the source directory to the target directory:
      hdfs dfs -mkdir [***DIRECTORY-NAME***]
      hdfs dfs -mv [***COMPLETE-PATH-TO-S3-BUCKET***]/[***SOURCE-DIRECTORY***] [***COMPLETE-PATH-TO-S3-BUCKET***]/[***TARGET-DIRECTORY***]
CDPD-48146: Error while browsing S3 buckets or ADLS containers from the left-assist panel
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs
When you might attempt to access the S3 buckets or ADLS containers from the left-assist panel in Data Explorer, a Failed to retrieve buckets: :1:0: syntax error message is displayed.
Access the S3 buckets or ADLS containers using the File Browser.
CDPD-42619: Unable to import a large CSV file from the local workstation
7.3.2, 7.3.1 and its SPs and CHFs, 7.2.18 and its SPs, 7.2.17 and its SPs
When you import a CSV file into Data Explorer from your workstation, an error might be displayed stating that you cannot import files of size more than 200 KB.
Upload the CSV file to an S3 bucket or ABFS container first, then import it into Data Explorer using the Importer.