Unsupported Interfaces and Features

You need to know the interfaces available in HDP or CDH platforms that are no longer supported in CDP. Some features you might have used are also unsupported.

Unsupported Interfaces

  • S3 for storing tables and LLAP (available in CDP Public Cloud only)
  • Hive CLI (replaced by Beeline)
  • WebHCat
  • Hcat CLI
  • SQL Standard Authorization
  • MapReduce execution engine (replaced by Tez)
  • Spark execution engine (replaced by Tez)
  • The spark thrift server

    Spark and Hive tables interoperate using the Hive Warehouse Connector.

  • Pig
  • Hive Indexes
  • Hive View and Tez View

    You can use Data Analytics Studio in lieu of Hive View.

Partially unsupported interfaces

Apache Hadoop Distributed Copy (DistCP) is not supported for copying Hive ACID tables. See link below.

Unsupported Features

CDP does not support the following features that were available in HDP and CDH platforms:

  • Replicate Hive ACID tables between CDP Private Cloud Base clusters using REPL commands

    You cannot use the REPL commands (REPL DUMP and REPL LOAD) to replicate Hive ACID table data between two CDP Private Cloud Base clusters.

  • CREATE TABLE that specifies a managed table location

    Do not use the LOCATION clause to create a managed table. Hive assigns a default location in the warehouse to managed tables.

  • CREATE INDEX

    Hive builds and stores indexes in ORC or Parquet within the main table, instead of a different table, automatically. Set hive.optimize.index.filter to enable use (not recommended--use materialized views instead). Existing indexes are preserved and migrated in Parquet or ORC to CDP during upgrade.

  • Hive metastore (HMS) high availablility (HA) load balancing

    You need to set up HMS HA as described in the documentation (see link below).

  • Local or Embedded Hive metastore server

    CDP does not support the use of a local or embedded Hive metastore setup.

Unsupported Connector Use

CDP does not support the Sqoop exports using the Hadoop jar command (the Java API) that Teradata documents. For more information, see Migrating data using Sqoop.