Behavioral Changes
Behavioral changes denote a marked change in behavior from the previously released version to this version of software. In HDP 2.6.3, behavioral changes affect the following Hadoop components.
Table 1.3. Behavioral Changes
Hortonworks Bug ID | Apache Component | Apache JIRA | Summary | Details |
---|---|---|---|---|
BUG-66121 | Hive | HIVE-14251 | Union All of different types resolves to incorrect data |
Scenario: UNION result handling Previous Behavior: queries in which union operator was used may have created an invalid output type.for example, the column type is ambigous in the following query: select cast(1 as int) union select cast(1 as string) Selecting the inappropriate type may cause the value to be changed
to New Behavior: The types are checked prior to execution; and the ambigous cases are rejected; FAILED: SemanticException Schema of both sides of union should match: Column _c0 is of type int on first table and type string on second table. Cannot tell the position of null AST. (state=42000,code=40000) The query should be clarified with explicit casts. |
BUG-80021 | Oozie | N/A | Modify references to yarn-client mode for Oozie Spark action |
Summary: Yarn-client mode for the Spark action is not supported. Component Affected: Spark action in Oozie Scenario: The yarn-client mode of the Spark action is no longer supported as of HDP 2.6.0. New Behavior: If you use yarn-client mode in Oozie or Falcon workflows, you must change the workflow to use yarn-cluster mode instead. Workflow Manager automatically converts imported workflows to yarn-cluster mode. |
BUG-85566 | Ranger | RANGER-1727 | Ranger allows user to change an external user's password with 'null' old password |
Summary: Ranger allows user to change an external user's password Scenario: External users password change should be performed at the external source (LDAP/AD etc). But having an ability to change the password via Ranger API (although it won't change the password in the source system) is not useful. Previous Behavior: API call to change external user's password was allowed (although this will not change the password in actual external source) New Behavior: Now, external user password change cannot be done via API. |
BUG-86663 | Atlas | ATLAS-2017 | Import API: update to make the new parameter to be optional |
Components Affected: Atlas, REST API end point api/admin/import Scenario: Users of Atlas REST API to import data will need to update for changes in content-type header Previous behavior: REST API to import data required content-type application/octet-stream New behavior: The REST API now requires content-type multipart/form-data. In addition, parameter request is made optional. For example: curl -g -X POST -u adminuser:password -H "Content-Type: multipart/form-data" -H "Cache-Control: no-cache" -F request=@importOptions.json -F data=@fileToBeImported.zip "http://localhost:21000/api/atlas/admin/import" |
BUG-87531 | HDFS | HDFS-10220 | A large number of expired leases can make namenode unresponsive and cause failover |
Component Affected: NameNode Scenario: Large number of expired leases can make namenode unresponsive and causes failover. Previous Behavior: NamenNode will try to recover all expired leases in a loop New Behavior: While releasing large number of leases NameNode will timeout after time configured by dfs.namenode.max-lock-hold-to-release-lease-ms to avoid holding lock for long times. |
BUG-88870 | HDFS | HDFS-10326 | Disable setting tcp socket send/receive buffers for write pipelines |
Component Affected: DataNode and the DFS client Previous Behavior: HDFS would explicitly set hardcoded values for TCP socket buffer sizes. New Behavior: The size of the TCP socket buffers are no longer hardcoded by default. Instead the OS now will automatically tune the size for the buffer. |
BUG-91290 | Hive, Ranger | N/A | Additional ranger hive policies required for INSERT OVERWRITE |
Scenario:Additional ranger hive policies required for INSERT OVERWRITE Previous behavior: Hive INSERT OVERWRITE queries succeed as usual. New behavior: Hive INSERT OVERWRITE queries are unexpectedly failing after upgrading to HDP-2.6.x with the error: Error while compiling statement: FAILED: HiveAccessControlException Permission denied: user jdoe does not have WRITE privilege on /tmp/*(state=42000,code=40000) As of HDP-2.6.0, Hive INSERT OVERWRITE queries require a Ranger URI policy to allow write operations, even if the user has write privilege granted through HDFS policy. Workaround/Expected Customer Action:
|
RMP-9153 | Zeppelin | ZEPPELIN-1515 | Support Zeppelin HDFS storage |
Previous Behavior: In releases of Zeppelin earlier than HDP-2.6.3, notebooks and configuration files were stored on the local disk of the Zeppelin server. New Behavior: With HDP-2.6.3+, the default storage is now in HDFS. Workaround/Expected Customer Action: When upgrading to HDP-2.6.3+ from versions earlier than HDP-2.6.3, perform the steps described in Enabling HDFS Storage for Zeppelin Notebooks and Configuration in HDP-2.6.3+. |