Command Line Interface
Set Up
The HCatalog command line interface (CLI) can be invoked as HIVE_HOME=hive_home hcat_home/bin/hcat where hive_home is the directory where Hive has been installed and hcat_home is the directory where HCatalog has been installed.
If you are using BigTop's rpms or debs you can invoke the CLI by doing /usr/bin/hcat.
HCatalog CLI
The HCatalog CLI supports these command line options:
Option |
Usage |
Description |
---|---|---|
-g |
hcat -g mygroup ... |
Tells HCatalog that the table which needs to be created must have group "mygroup". |
-p |
hcat -p rwxr-xr-x ... |
Tells HCatalog that the table which needs to be created must have permissions "rwxr-xr-x". |
-f |
hcat -f myscript.hcatalog ... |
Tells HCatalog that myscript.hcatalog is a file containing DDL commands to execute. |
-e |
hcat -e 'create table mytable(a int);' ... |
Tells HCatalog to treat the following string as a DDL command and execute it. |
-D |
hcat -Dkey=value ... |
Passes the key-value pair to HCatalog as a Java System Property. |
hcat |
Prints a usage message. |
Note the following:
- The -g and -p options are not mandatory.
- Only one -e or -f option can be provided, not both.
- The order of options is immaterial; you can specify the options in any order.
If no option is provided, then a usage message is printed:
Usage: hcat { -e "<query>" | -f <filepath> } [-g <group>] [-p <perms>] [-D<name>=<value>]
Owner Permissions
When using the HCatalog CLI, you cannot specify a permission string without read permissions for owner, such as -wxrwxr-x, because the string begins with "-". If such a permission setting is desired, you can use the octal version instead, which in this case would be 375. Also, any other kind of permission string where the owner has read permissions (for example r-x------ or r--r--r--) will work fine.
HCatalog DDL
HCatalog supports all Hive Data Definition Language except those operations that require running a MapReduce job. For commands that are supported, any variances are noted below.
HCatalog does not support the following Hive DDL commands:
- IMPORT FROM ...
- EXPORT TABLE
- CREATE TABLE ... AS SELECT
- ALTER TABLE ... REBUILD
- ALTER TABLE ... CONCATENATE
- ANALYZE TABLE ... COMPUTE STATISTICS
- ALTER TABLE ARCHIVE/UNARCHIVE PARTITION
Create/Drop/Alter Table
CREATE TABLE
If you create a table with a CLUSTERED BY clause you will not be able to write to it with Pig or MapReduce. This is because they do not understand how to partition the table, so attempting to write to it would cause data corruption.
CREATE TABLE AS SELECT
Not supported. Throws an exception with the message "Operation Not Supported".
DROP TABLE
Supported. Behavior the same as Hive.
ALTER TABLE
Supported except for the REBUILD and CONCATENATE options. Behavior the same as Hive.
Create/Drop/Alter View
Note: Pig and MapReduce cannot read from or write to views.
CREATE VIEW
Supported. Behavior same as Hive.
DROP VIEW
Supported. Behavior same as Hive.
ALTER VIEW
Supported. Behavior same as Hive.
Show/Describe
SHOW TABLES
Supported. Behavior same as Hive.
SHOW PARTITIONS
Not supported. Throws an exception with message "Operation Not Supported".
SHOW FUNCTIONS
Supported. Behavior same as Hive.
DESCRIBE
Supported. Behavior same as Hive.
Create/Drop Index
CREATE and DROP INDEX operations are supported.
Note: Pig and MapReduce cannot write to a table that has auto rebuild on, because Pig and MapReduce do not know how to rebuild the index.
Create/Drop Function
CREATE and DROP FUNCTION operations are supported, but created functions must still be registered in Pig and placed in CLASSPATH for MapReduce.
"dfs" Command and "set" Command
Supported. Behavior same as Hive.
Other Commands
Any command not listed above is NOT supported and throws an exception with the message "Operation Not Supported".
CLI Errors
Authentication
If a failure results in a message like "2010-11-03 16:17:28,225 WARN hive.metastore ... - Unable to connect metastore with URI thrift://..." in /tmp/<username>/hive.log, then make sure you have run "kinit <username>@FOO.COM" to get a Kerberos ticket and to be able to authenticate to the HCatalog server. |
Error Log
If other errors occur while using the HCatalog CLI, more detailed messages are written to /tmp/<username>/hive.log.