Known Issues in Streaming Analytics

Learn about the known issues in Streaming Analytics clusters, the impact or changes to the functionality, and the workaround.

SQL Stream Builder

7.2.16.1
FLINK-18027: ROW value constructor cannot deal with complex expressions
When querying data from a table or a view with a ROW() function an exception is thrown due to a Calcite parsing issue. For example, the following query will return an error:
CREATE VIEW example AS SELECT col1, ROW(col2) FROM table;
SELECT * FROM example;
Add a second SELECT layer to the SQL query as shown in the following example:
CREATE VIEW example AS SELECT col1, ROW(col2) FROM (SELECT col1, col2 FROM table);
SELECT * FROM example;
CSA-4541: Confluent Kafka cannot be validated
When using Confluent Kafka with SSB, the data source cannot be validated with SASL/SSL security method as the default truststore path is configured for CDP environments, and there is no option for custom parameters to be configured.
None
CSA-4426: Kafka Data Source name accepts spaces
Kafka Data Source can be validated and created with spaces in the Data Source name, but this results in errors as spaces are not valid characters based on the naming convention.
None
CSA-4464: CSA parcel is built with an interim CDP build
The CSA parcel is built using an interim CDP build and not with a build that corresponds to a release version. This can cause errors with components that have dependency to Flink.
If a Flink component transitively depends on some CDP related module and it is not accessible publicly, the dependency can be excluded. In case the project also depends on the excluded module, the publicly available version of the dependency can be added to Flink. For example, this could happen with kafka-clients, which is pulled in by flink-connector-kafka:
...
<dependency>
	<groupId>org.apache.flink</groupId>
	<artifactId>flink-connector-kafka</artifactId>
	<version>${flink.version}</version>
	<exclusions>
		<exclusion>
			<groupId>org.apache.kafka</groupId>
			<artifactId>kafka-clients</artifactId>
		</exclusion>
	</exclusions>
</dependency>

<dependency>
	<groupId>org.apache.kafka</groupId>
	<artifactId>kafka-clients</artifactId>
	<version>3.1.2.7.2.16.0-287</version>
</dependency>
...
CSA-4426: Kafka Data Source name accepts spaces
Kafka Data Source can be validated and created with spaces in the Data Source name, but this results in errors as spaces are not valid characters based on the naming convention.
None
CSA-4412: Cannot delete Materialized View endpoint when using dynamic parameters
Materialized View endpoints cannot be deleted if dynamic parameters were set for them.
None
7.2.16.0
CSA-4650: Inconsistent sidebar collapse behavior
The sidebar is collapsed inconsistently on the homepage of Streaming SQL Console when opening a project.
None
CSA-4643: flink-yarn-session is ignoring command line parameters
When adding parameters to the Flink session using flink-yarn-session -d in command line, the parameters are not applied to the session.
None
CSA-4548: Files cannot be uploaded through Swagger
REST API endpoints that take multipart requests, such as the uploading artifact endpoint, result in error.
None
CSA-4428: upsert-kafka template has an invalid property
The upsert Kafka template contains an invalid format property, which causes the table creation to fail.
None
CSA-4426: Kafka Data Source name accepts spaces
Kafka Data Source can be validated and created with spaces in the Data Source name, but this results in errors as spaces are not valid characters based on the naming convention.
None
CSA-4413: Missing AWS library
Due to the missing AWS library, SQL Stream Builder and Flink cannot be used in a RAZ-enabled environment.
None
CSA-4412: Cannot delete Materialized View endpoint when using dynamic parameters
Materialized View endpoints cannot be deleted if dynamic parameters were set for them.
None
CSA-4400: Cannot delete invalid catalog
A catalog can be created without a catalog service, but deleting the invalid catalog fails as it is not registered in Flink without a service.
None
CSA-4390: Missing flink-shaded-zookeeper-3 artifacts
The flink-shaded-zookeeper-3 artifacts are missing the public artifact repository.
Exclude the flink-shaded-zookeeper-3 dependency from the modules that would try to pull it in transitively. The following modules are relevant for the Flink application projects:
  • flink-streaming-java
  • flink-streaming-scala_2.12
  • flink-table-api-java-bridge
  • flink-table-api-scala-bridge_2.12
  • flink-table-api-java-uber
  • flink-clients, flink-statebackend-*
  • flink-connector-*
  • flink-sql-connector-*
  • flink-test-utils
  • flink-connector-test-utils
  • flink-table-test-utils
CSA-4378: Validating Kafka Data Source fails without security settings
Kafka Data Source validation fails in a secure environment if the truststore details are not provided.
None
CSA-4377: Improve masking information of credentials
Credential information might be exposed on different configuration pages.
None
CSA-4372: Allow Deletion is not synchronized with project
When configuring a project with Allow Deletion setting and exported to a git repository, the Allow Deletion is not synchronized.
None
CSA-4370: Virtual tables imported from a schema in a Schema Catalog fail to describe correctly
Describing fails for a Virtual Table from a schema in a Schema Registry Catalog when viewing the table DDL.
None
CSA-4363: Production mode does not use clean session variables
When setting a variable in production mode, the job fails with an error as the variables set in production mode will not be used regardless of resetting the value of variable.
None
CSA-4360: Activating an environment does not reload catalogs
When activating an environment file in which variables are used for registered catalogs, the catalogs do not show up only after revalidation or after registering them again.
None
CSA-4356: Read-only jobs can be deleted
Running and read-only jobs can be deleted even though a member of a project does not have the permissions to stop and restart them.
None
CSA-4355: Virtual Table tab keeps loading for views
The Virtual Table tab is loading for an undefined time when opening a created view.
None
CSA-4354: Kudu lookup join fails
When using a Kudu lookup join and adding a WHERE clause with a string, the SSB job fails.
None
CSA-4353: Editing DDL Kafka tables opens wizard
When editing a Kafka table that was created using DDL, the configuration window appears as if the table was created using the Kafka Table wizard. This results in showing incorrect table information.
None
CSA-4352: Schema is not shown in table view
If a virtual table is opened in the table view, the schema is not shown and the information is displayed incorrectly.
None
CSA-4337: Materialized View link throws error
After importing a project and running the existing SQL job, the Materialized View query link does not work and the following error is shown:
"No data available yet. Verify that your Materialized View is configured, and the streaming job is running, and refresh page."
None
CSA-4333: Use Kafka Timestamps switch reflects invalid value
After creating a Kafka Virtual Table and disabling the “Use Kafka Timestamps” configuration, the table is created successfully according to the setting, but when viewing the DDL of the table, it shows the configuration as enabled.
None
CSA-4030: Webhook sending fails when webhook template is empty string
When creating a webhook table with a custom template, the webhook template will be saved as an empty string, which results in webhook sending failure.
None
CSA-3886: Session timeout configuration is not enforced
When setting the SSB Session Timeout configuration in Cloudera Manager, the set time is not enforced and SSB continues to run indefinitely.
None
CSA-3867: UDF case sensitivity
Due to a case sensitivity mismatch, User-Defined Functions (UDFs) can be created with uppercase, and be called by lowercase, but not vice-versa. This results in error when executing jobs as the UDFs are saved differently in the database.
None
CSA-3498: SSB fails with IllegalStateException
SSB fails to work and throws an IllegalStateException when running a job due to sample websocket session timeout.
None

Flink

7.2.16.0, 7.2.16.1
FLINK-18027: ROW value constructor cannot deal with complex expressions
When querying data from a table or a view with a ROW() function an exception is thrown due to a Calcite parsing issue. For example, the following query will return an error:
CREATE VIEW example AS SELECT col1, ROW(col2) FROM table;
SELECT * FROM example;
Add a second SELECT layer to the SQL query as shown in the following example:
CREATE VIEW example AS SELECT col1, ROW(col2) FROM (SELECT col1, col2 FROM table);
SELECT * FROM example;
In Cloudera Streaming Analytics, the following SQL API features are in preview:
  • Match recognize
  • Top-N
  • Stream-Table join (without rowtime input)