Hortonworks Docs
»
DataFlow 3.3.0
»
Installing HDF Services on an Existing HDP Cluster
Installing HDF Services on an Existing HDP Cluster
Also available as:
Upgrade Ambari and HDP
Installing Databases
Installing MySQL
Configuring SAM and Schema Registry Metadata Stores in MySQL
Configuring Druid and Superset Metadata Stores in MySQL
Install Postgres
Configure Postgres to Allow Remote Connections
Configure SAM and Schema Registry Metadata Stores in Postgres
Configure Druid and Superset Metadata Stores in Postgres
Specifying an Oracle Database to Use with SAM and Schema Registry
Switching to an Oracle Database After Installation
Installing the HDF Management Pack
Update the HDF Base URL
Add HDF Services to an HDP Cluster
Configure HDF Components
Configure Schema Registry
Configure SAM
Configuring SAM log search and event sampling
Configure NiFi
Configure NiFi for Atlas Integration
Configure Kafka
Configure Storm
Configure Log Search
Deploy the Cluster Services
Access the UI for Deployed Services
Configuring Schema Registry and SAM for High Availability
Configuring SAM for High Availability
Configuring Schema Registry for High Availability
Deploy the Cluster Services
Finish the wizard and deploy the cluster. After the cluster has been deployed, some services might fail in starting. If this is the case, start those services individually.
Parent topic:
Configure HDF Components
© 2012–2019, Hortonworks, Inc.
Document licensed under the
Creative Commons Attribution ShareAlike 4.0 License
.
Hortonworks.com
|
Documentation
|
Support
|
Community