PutClouderaORC 2.3.0.4.10.0.0-147

Bundle
com.cloudera | nifi-cdf-hive-nar
Description
Reads records from an incoming FlowFile using the provided Record Reader, and writes those records to a ORC file in the location/filesystem specified in the configuration.
Tags
HDFS, ORC, filesystem, hadoop, put, record, restricted
Input Requirement
REQUIRED
Supports Sensitive Dynamic Properties
false
Properties
Restrictions
Required Permission Explanation
write distributed filesystem Provides operator the ability to write to any file that NiFi has access to in HDFS or the local filesystem.
Relationships
Name Description
retry Flow Files that could not be processed due to issues that can be retried are transferred to this relationship
success Flow Files that have been successfully processed are transferred to this relationship
failure Flow Files that could not be processed due to issue that cannot be retried are transferred to this relationship
Reads Attributes
Name Description
filename The name of the file to write comes from the value of this attribute.
Writes Attributes
Name Description
filename The name of the file is stored in this attribute.
absolute.hdfs.path The absolute path to the file is stored in this attribute.
hadoop.file.url The hadoop url for the file is stored in this attribute.
record.count The number of records written to the ORC file
hive.ddl Creates a partial Hive DDL statement for creating an external table in Hive from the destination folder. This can be used in ReplaceText for setting the content to the DDL. To make it valid DDL, add "LOCATION '<path_to_orc_file_in_hdfs>'", where the path is the directory that contains this ORC file on HDFS. For example, this processor can send flow files downstream to ReplaceText to set the content to this DDL (plus the LOCATION clause as described), then to PutHiveQL processor to create the table if it doesn't exist.