Write FlowFile data to Hadoop Distributed File System (HDFS)
hadoop, HCFS, HDFS, put, copy, filesystem
In the list below, the names of required properties appear in bold. Any other properties (not in bold) are considered optional. The table also indicates any default values, and whether a property supports the NiFi Expression Language.
Display Name | API Name | Default Value | Allowable Values | Description |
---|---|---|---|---|
Hadoop Configuration Resources | Hadoop Configuration Resources | A file or comma separated list of files which contains the Hadoop file system configuration. Without this, Hadoop will search the classpath for a 'core-site.xml' and 'hdfs-site.xml' file or will revert to a default configuration. To use swebhdfs, see 'Additional Details' section of PutHDFS's documentation. This property expects a comma-separated list of file resources. Supports Expression Language: true (will be evaluated using variable registry only) | ||
Kerberos Credentials Service | kerberos-credentials-service | Controller Service API: KerberosCredentialsService Implementation: KeytabCredentialsService | Specifies the Kerberos Credentials Controller Service that should be used for authenticating with Kerberos | |
Kerberos User Service | kerberos-user-service | Controller Service API: KerberosUserService Implementations: KerberosTicketCacheUserService KerberosKeytabUserService KerberosPasswordUserService | Specifies the Kerberos User Controller Service that should be used for authenticating with Kerberos | |
Kerberos Principal | Kerberos Principal | Kerberos principal to authenticate as. Requires nifi.kerberos.krb5.file to be set in your nifi.properties Supports Expression Language: true (will be evaluated using variable registry only) | ||
Kerberos Keytab | Kerberos Keytab | Kerberos keytab associated with the principal. Requires nifi.kerberos.krb5.file to be set in your nifi.properties This property requires exactly one file to be provided.. Supports Expression Language: true (will be evaluated using variable registry only) | ||
Kerberos Password | Kerberos Password | Kerberos password associated with the principal. Sensitive Property: true | ||
Kerberos Relogin Period | Kerberos Relogin Period | 4 hours | Period of time which should pass before attempting a kerberos relogin.
This property has been deprecated, and has no effect on processing. Relogins now occur automatically. Supports Expression Language: true (will be evaluated using variable registry only) | |
Additional Classpath Resources | Additional Classpath Resources | A comma-separated list of paths to files and/or directories that will be added to the classpath and used for loading native libraries. When specifying a directory, all files with in the directory will be added to the classpath, but further sub-directories will not be included. This property expects a comma-separated list of resources. Each of the resources may be of any of the following types: file, directory. | ||
Directory | Directory | The parent HDFS directory to which files should be written. The directory will be created if it doesn't exist. Supports Expression Language: true (will be evaluated using flow file attributes and variable registry) | ||
Conflict Resolution Strategy | Conflict Resolution Strategy | fail |
| Indicates what should happen when a file with the same name already exists in the output directory |
Writing Strategy | writing-strategy | Write and rename |
| Defines the approach for writing the FlowFile data. |
Block Size | Block Size | Size of each block as written to HDFS. This overrides the Hadoop Configuration | ||
IO Buffer Size | IO Buffer Size | Amount of memory to use to buffer file contents during IO. This overrides the Hadoop Configuration | ||
Replication | Replication | Number of times that HDFS will replicate each file. This overrides the Hadoop Configuration | ||
Permissions umask | Permissions umask | A umask represented as an octal number which determines the permissions of files written to HDFS. This overrides the Hadoop property "fs.permissions.umask-mode". If this property and "fs.permissions.umask-mode" are undefined, the Hadoop default "022" will be used. If the PutHDFS target folder has a default ACL defined, the umask property is ignored by HDFS. | ||
Remote Owner | Remote Owner | Changes the owner of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change owner Supports Expression Language: true (will be evaluated using flow file attributes and variable registry) | ||
Remote Group | Remote Group | Changes the group of the HDFS file to this value after it is written. This only works if NiFi is running as a user that has HDFS super user privilege to change group Supports Expression Language: true (will be evaluated using flow file attributes and variable registry) | ||
Compression codec | Compression codec | NONE |
| No Description Provided. |
Ignore Locality | Ignore Locality | false |
| Directs the HDFS system to ignore locality rules so that data is distributed randomly throughout the cluster |
Name | Description |
---|---|
success | Files that have been successfully written to HDFS are transferred to this relationship |
failure | Files that could not be written to HDFS for some reason are transferred to this relationship |
Name | Description |
---|---|
filename | The name of the file written to HDFS comes from the value of this attribute. |
Name | Description |
---|---|
filename | The name of the file written to HDFS is stored in this attribute. |
absolute.hdfs.path | The absolute path to the file on HDFS is stored in this attribute. |
hadoop.file.url | The hadoop url for the file is stored in this attribute. |
target.dir.created | The result(true/false) indicates if the folder is created by the processor. |
Required Permission | Explanation |
---|---|
write distributed filesystem | Provides operator the ability to delete any file that NiFi has access to in HDFS or the local filesystem. |