Known Issues
The CCP 2.0.1 release has the following known issues:
- Installing Spark 2 fails
- Workaround: Before attempting to install Spark2 with Ambari, install the MySQL Connector manually and ensure Ambari can find it. This needs to be run on the node hosting Ambari.
- Kerberos authentication against HDFS from Metron's Storm topologies can fail.
-
If this issue occurs, the Storm worker is unable to present a valid Kerberos ticket to authenticate against HDFS. This impacts the Enrichment and Batch Indexing topologies, each of which interact with HDFS.
- During CCP installation, some versions of Zeppelin might fail to install.
- If the Zeppelin notebooks are not installed, import the Apache Zeppelin Notebook manually.
- The Kerberization process might lock solr directories.
- If this occurs you will see the following message in the logs:
is locked (lockType=hdfs). Throwing exception.
and you will not see Solr alerts in the Alerts UI. If this issue occurs, remove thewrite.lock
file located at/solr/bro/core_node1/data-index/write.lock
or, in Ambari, navigate to Solr > config > Advanced solr-hdfs and check the Delete write.lock files on HDFS checkbox. After you have deleted thewrite.lock
file, restart Solr. - When running a large sized PCAP query, the REST API can die silently if the result set exceeds the memory available to the REST server.
- On Kerberized clusters Storm rebalances can fail to correctly distribute tickets.
- This can be resolved by running
storm upload-credentials
against each topology.