Known issues in Hive Virtual Warehouses
This topic describes the Hive Virtual Warehouse known issues for Cloudera Data Warehouse (CDW) Private Cloud.
Known issues identified in 1.5.4
- DWX-18505: Resource template changes back to Reduced resources while creating a Virtual Warehouse
- On the Create Virtual Warehouse modal,
suppose you selected Default resources from the
Resource Template drop-down menu. When you wait for a few
seconds or click elsewhere on the modal, you see that the resource template changes back
to
Reduced resources
. However, when you create the Virtual Warehouse, the Virtual Warehouse is created with the resource template you initially selected. This is a CDW UI defect you encounter on environments that are activated using the low resource mode. - DWX-18445: Virtual Warehouses get stuck in the deleting state
- You may intermittently notice that a Virtual Warehouses gets
stuck in the
deleting
state when you delete a Virtual Warehouse in CDW. This could be due to a failing diagnostic job that is in anInit:0/1
state in the namespace.
Known issues identified in 1.5.2
- DWX-16989: Hive query running on Iceberg table fails randomly
- Suppose you have disabled the auto-suspend option for a Hive
Virtual Warehouse or if the Virtual Warehouse is under continuous load and hence it
cannot be stopped by the auto-suspend option. In this situation, using the Iceberg table
format may cause the following exceptions to appear in the query coordinator log along
with the submitted queries that have
failed:
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken): token (token for hive: HDFS_DELEGATION_TOKEN owner=hive/dwx-env-host-1.cdp.local@EXAMPLE.CLOUDERA.COM, renewer=hive, realUser=, issueDate=1709813340891, maxDate=1710418140891, sequenceNumber=19784486, masterKeyId=52) is expired, current time: 2024-03-08 04:09:32,835-0800 expected renewal time: 2024-03-08 04:09:00,891-0800
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken): token (token for hive: HDFS_DELEGATION_TOKEN owner=hive/dwx-env-host-1.cdp.local@EXAMPLE.CLOUDERA.COM, renewer=hive, realUser=, issueDate=1699855596578, maxDate=1700460396578, sequenceNumber=16863242, masterKeyId=39) can't be found in cache
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken): token (OzoneToken owner=hive/dwx-env-ewxf6g-env.cdp.local@ROOT.EXAMPLE.SITE, renewer=hive, realUser=, issueDate=2024-03-19T21:49:31.033Z, maxDate=2024-03-19T21:50:31.033Z, sequenceNumber=72, masterKeyId=1, strToSign=null, signature=null, awsAccessKeyId=null, omServiceId=ozone1710521984, omCertSerialId=11) is expired, current time: 2024-03-19 21:51:34,293+0000 expected renewal time: 2024-03-19 21:51:31,033+0000
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken): token (OzoneToken owner=hive/dwx-env-azt1gg-env.cdp.local@ROOT.EXAMPLE.SITE, renewer=hive, realUser=, issueDate=2024-04-09T16:04:12.889Z, maxDate=2024-04-09T17:04:12.889Z, sequenceNumber=29, masterKeyId=1, strToSign=null, signature=null, awsAccessKeyId=null, omServiceId=ozone1711550158, omCertSerialId=2597525731772327) can't be found in cache
This happens because the HDFS delegation tokens are not renewed when using the Iceberg table format. After the existing HDFS delegation tokens expire, Hive query coordinator (TEZ App Master) cannot access the tables on the file system during the query planning phase. The problem is independent of the file system--Ozone FS or Hadoop FS. The error only occurs after the HDFS delegation tokens have expired. By default, the delegation tokens expire in one day. However, you can modify the expiration time on the CDP Base cluster.
The problem does not occur if the query coordinator pods in the Hive Virtual Warehouse are stopped manually or by using the auto-suspend functionality within the token expiration period.
Known issues identified in 1.5.1
- DWX-15480: Hive queries fail with FILE_NOT_FOUND error
- ACID directory cache may become outdated in Tez AMs in case of ACID tables that change often, possibly leading to different errors with the same root cause: "split generation works from cache pointing to non-existing files". And you may see the following error in the diagnostic bundles and query logs: FILE_NOT_FOUND: Unable to get file status.
- DWX-15287: Drop database query for Hive fails with Invalid ACL Exception
- You may see the following error in a Hue or beeline session when
running DROP DATABASE, DROP TABLE, or ALTER TABLE DROP PARTITION operations on a Hive
Virtual Warehouse that is in Stopped state:
"org.apache.zookeeper.KeeperException$InvalidACLException:
KeeperErrorCode = InvalidACL for /llap-sasl/user-hive".The exception is caused because the Hive Virtual Warehouse tries to evict the cache in the LLAP executors, but the compute pods of the stopped warehouse are no longer running.