Storage
Cloudera Runtime provides different types of storage components that you can use depending on your data requirements. Apache Hadoop HDFS is a distributed file system for storing large volumes of data. Apache Kudu completes Apache Hadoop’s storage layer, enabling fast analytics on fast data.
- Managing Data Storage
- Provides information about optimizing data storage, APIs and services for accessing data, and managing data across clusters.
- Configuring Data Protection
- Provides information about configuring data protection on a Hadoop cluster.
- Configuring HDFS ACLs
- Describes the procedure to configure Access Control Lists (ACLs) on Apache Hadoop HDFS.
- Configuring Fault Tolerance
- Describes the procedure to configure HDFS high availability on a cluster.
- Administering Apache Kudu
- Describes common administrative tasks and Apache Kudu workflows.
- Developing Applications with Apache Kudu
- Provides reference examples to use C++ and Java client APIs to develop apps using Apache Kudu.
- Using Apache Impala with Apache Kudu
- Provides information about how to use use Kudu as a storage for Apache Impala.
- Accessing Cloud Data
- Describes information about the configuration parameters used to access data stored in the cloud.