Hortonworks Docs
»
DataFlow 3.3.0
»
Installing HDF Services on a New HDP Cluster
Installing HDF Services on a New HDP Cluster
Also available as:
Installing Ambari
Installing Databases
Installing MySQL
Configuring SAM and Schema Registry Metadata Stores in MySQL
Configuring Druid and Superset Metadata Stores in MySQL
Install Postgres
Configure Postgres to Allow Remote Connections
Configure SAM and Schema Registry Metadata Stores in Postgres
Configure Druid and Superset Metadata Stores in Postgres
Specifying an Oracle Database to Use with SAM and Schema Registry
Switching to an Oracle Database After Installation
Deploying an HDP Cluster Using Ambari
Installing an HDP Cluster
Customize Druid Services
Configure Superset
Deploy the Cluster Services using Ambari
Access the Stream Insight Superset UI
Installing the HDF Management Pack
Update the HDF Base URL
Add HDF Services to an HDP Cluster
Configure HDF Components
Configure Schema Registry
Configure SAM
Configuring SAM log search and event sampling
Configure NiFi
Configure NiFi for Atlas Integration
Configure Kafka
Configure Storm
Configure Log Search
Deploy the Cluster Services
Access the UI for Deployed Services
Configuring Schema Registry and SAM for High Availability
Configuring SAM for High Availability
Configuring Schema Registry for High Availability
Deploy the Cluster Services
Finish the wizard and deploy the cluster. After the cluster has been deployed, some services might fail in starting. If this is the case, start those services individually.
Parent topic:
Configure HDF Components
© 2012–2019, Hortonworks, Inc.
Document licensed under the
Creative Commons Attribution ShareAlike 4.0 License
.
Hortonworks.com
|
Documentation
|
Support
|
Community