Homepage
/
Cloudera Streaming Analytics
1.6.2
(Private Cloud)
Search Documentation
▶︎
Cloudera
Reference Architectures
▶︎
Cloudera Public Cloud
Getting Started
Patterns
Preview Features
Data Catalog
Data Engineering
DataFlow
Data Hub
Data Warehouse
Data Warehouse Runtime
Cloudera AI
Management Console
Operational Database
Replication Manager
DataFlow for Data Hub
Runtime
▼
Cloudera Private Cloud
Data Services
Getting Started
Cloudera Manager
Management Console
Data Engineering
Data Warehouse
CDW Runtime
Machine Learning
Base
Getting Started
Runtime & Cloudera Manager
Upgrade
Flow Management
Streaming Analytics
▶︎
Cloudera Manager
Cloudera Manager
▶︎
Applications
Cloudera Streaming Community Edition
Data Science Workbench
Data Visualization
Edge Management
Observability SaaS
Observability on premises
Workload XM On-Prem
▶︎
Legacy
Cloudera Enterprise
Flow Management
Stream Processing
HDP
HDF
Streams Messaging Manager
Streams Replication Manager
▶︎
Data Services
Getting Started
Cloudera Manager
Management Console
Data Engineering
Data Warehouse
CDW Runtime
Machine Learning
Base
Getting Started
Runtime & Cloudera Manager
Upgrade
Flow Management
Streaming Analytics
«
Filter topics
Cloudera Streaming Analytics
▶︎
Release Notes
What's new in Cloudera Streaming Analytics
Fixed issues
Component support
Known issues and limitations
Unsupported features
▶︎
Download
Download location of CSA
▶︎
Apache Flink Overview
Streaming Analytics in Cloudera
What is Apache Flink?
Core Features of Flink
▶︎
SQL Stream Builder Overview
Introduction to SQL Stream Builder
Key features of SSB
SQL Stream Builder architecture
▶︎
Quick Start
▶︎
DataStream Applications
Quick Start for Flink
Flink Project Template
▶︎
SQL Stream Applications
Quick Start for SSB
▶︎
Deployment
Deployment scenarios
Cluster service layout with Flink
Cluster service layout with SSB
▶︎
Installation & Upgrade
▶︎
Support Matrix
System Requirements
Flink API Support
Default ports for Flink and SSB
▶︎
Installation
Installing CSD and parcel
▶︎
Adding Flink as a Service
Setting up your HDFS Home directory
Setting the Java executable for the Flink client
▶︎
Configuring databases for SSB
Setting up MySQL/MariaDB database for SSB
Setting up PostgreSQL database for SSB
Setting up Oracle database for SSB
Adding SSB as a service
▶︎
Upgrade
▶︎
Upgrading Flink service
Creating new Flink application versions
Stopping Flink applications
Upgrading CSA artifacts for Flink service
Resuming Flink applications
▶︎
Upgrading SQL Stream Builder service
Stopping SQL Stream jobs
Upgrading CSA artifacts for SQL Stream Builder service
Resuming SQL jobs
▶︎
Migration
Migrating Flink service to a different host
Migrating SQL jobs
▼
Security
▶︎
Securing Apache Flink
Authentication and encryption for Flink
▶︎
Enabling security for Apache Flink
Configuring custom Kerberos principal for Apache Flink
Enabling SPNEGO authentication for Flink Dashboard
▶︎
Enabling Knox authentication for Flink Dashboard
Enabling Knox Auto Discovery for Flink
Enabling Knox manually for Flink Dashboard
Accessing the Flink Dashboard through Knox
Configuring Ranger policies for Flink
Securing Apache Flink jobs
Using EncryptTool for Flink properties
▼
Securing SQL Stream Builder
▶︎
Authentication in SSB
▶︎
Enabling Kerberos authentication
Configuring custom Kerberos principal for SQL Stream Builder
▶︎
Enabling Knox authentication
Enabling Knox for SSB in CDP Private Cloud Base 7.1.6
Enabling Knox for SSB in CDP Private Cloud Base 7.1.7
Accessing the Streaming SQL Console through Knox
Uploading and unlocking your keytab in SSB
▶︎
Encryption in SSB
Enabling TLS for database connection
Configuring Ranger policies for SSB
Managing teams in Streaming SQL Console
▶︎
Using SQL Stream Builder
▶︎
Getting Started
Running a simple SQL job
▶︎
Streaming SQL Console
▶︎
Using the Streaming SQL Console
▶︎
Console Page
Compose Tab
Tables Tab
Functions Tab
History Tab
SQL Jobs Tab
Data Providers Page
Materialized Views Page
▶︎
Data Providers
▶︎
Registering Data Providers in SSB
Adding Kafka as Data Provider
Adding Schema Registry as Catalog
Adding Kudu as Catalog
Adding Hive as Catalog
Adding Custom Catalogs
Managing registered Data Providers
▶︎
Connectors
▶︎
Connector support in SSB
Kafka connectors
CDC connectors
JDBC connector
Filesystem connector
Datagen connector
Faker connector
Blackhole connector
▶︎
Managing connectors and data formats
▶︎
Adding new connectors
SSB fails with missing JAR file error
Adding data formats
▶︎
Creating Tables
Concept of tables in SSB
▶︎
Creating Kafka tables
Creating Kafka tables using Console wizard
Creating Kafka tables using Templates
▶︎
Configuring Kafka tables
Schema tab
Event Time tab
Transformations tab
Properties tab
Deserialization tab
Assigning Kafka keys in streaming queries
Performance & Scalability
Creating Flink tables using Templates
Creating Webhook tables
Managing time in SSB
▶︎
Flink SQL
▶︎
Flink SQL Overview
Flink DDL
Flink DML
Flink Queries
Other supported statements
Data Types
Dynamic SQL Hints
SQL Examples
▶︎
Enriching streaming data with join
Joining streaming and bounded tables
Example: joining Kafka and Kudu tables
▶︎
Job Lifecycle
Running SQL Stream jobs
▶︎
Configuring SQL job settings
Adjusting logging configuration in Advanced Settings
Stopping, restarting and editing SQL jobs
Sampling data for a running job
▶︎
Managing session for SQL jobs
Executing SQL jobs in production mode
▶︎
REST API
Using SQL Stream Builder REST API
▶︎
Materialized Views
Introduction to Materialized Views
Creating Materialized Views
Using Dynamic Materialized View Endpoints
Using SQL Stream Builder with Cloudera Data Visualization
▶︎
Input Transforms
Creating Input Transforms
▶︎
User Defined Functions
Creating User Defined Functions
Developing JavaScript functions
Adding Java to the Functions language option
Using System Functions
▶︎
Monitoring
Monitoring SQL Stream jobs
▶︎
Governance
SSB metadata collection using Atlas
▶︎
Using Apache Flink
▶︎
Getting Started
Running a simple Flink application
▶︎
Application Development
▶︎
Flink application structure
Source, operator and sink in DataStream API
Flink application example
Testing and validating Flink applications
▶︎
Configuring Flink applications
Setting parallelism and max parallelism
Configuring Flink application resources
Configuring RocksDB state backend
Enabling checkpoints for Flink applications
▶︎
DataStream Connectors
▶︎
HBase sink with Flink
Creating and configuring the HBaseSinkFunction
▶︎
Kafka with Flink
▶︎
Schema Registry with Flink
ClouderaRegistryKafkaSerializationSchema
ClouderaRegistryKafkaDeserializationSchema
Kafka Metrics Reporter
Kudu with Flink
▶︎
Job Lifecycle
Running a Flink job
Using Flink CLI
Enabling savepoints for Flink applications
▶︎
Monitoring
Flink Dashboard
Streams Messaging Manager integration
Enabling Flink DEBUG logging
▶︎
SQL and Table API
▶︎
Flink SQL and Table API
SQL and Table API supported features
▶︎
DataStream API interoperability
Converting DataStreams to Tables
Converting Tables to DataStreams
Supported data types
▶︎
SQL catalogs for Flink
Hive catalog
Kudu catalog
Schema Registry catalog
▶︎
SQL connectors for Flink
Kafka connector
▶︎
Data types for Kafka connector
JSON format
CSV format
▶︎
Avro format
Supported basic data types
Schema Registry Avro format
▶︎
SQL Statements in Flink
CREATE Statements
DROP Statements
ALTER Statements
INSERT Statements
SQL Queries in Flink
▶︎
Governance
▶︎
Flink metadata collection using Atlas
Atlas entities in Flink metadata collection
Creating Atlas entity type definitions for Flink
Verifying metadata collection
▶︎
Reference
Flink Terminology
Cloudera Flink Tutorials
▶︎
Storm Flink Migration
▶︎
Comparing Storm and Flink
Conceptual differences
Differences in architecture
Differences in data distribution
Migrating from Storm to Flink
SQL Stream Builder REST API Reference
Accessing the Flink Dashboard through Knox
Accessing the Streaming SQL Console through Knox
Adding Custom Catalogs
Adding data formats
Adding Flink as a Service
Adding Hive as Catalog
Adding Java to the Functions language option
Adding Kafka as Data Provider
Adding Kudu as Catalog
Adding new connectors
Adding Schema Registry as Catalog
Adding SSB as a service
Adjusting logging configuration in Advanced Settings
ALTER Statements
Apache Flink Overview
Application Development
Assigning Kafka keys in streaming queries
Atlas entities in Flink metadata collection
Authentication and encryption for Flink
Authentication in SSB
Avro format
Blackhole connector
CDC connectors
Cloudera Flink Tutorials
Cloudera Streaming Analytics
ClouderaRegistryKafkaDeserializationSchema
ClouderaRegistryKafkaSerializationSchema
Cluster service layout with Flink
Cluster service layout with SSB
Comparing Storm and Flink
Component support
Compose Tab
Concept of tables in SSB
Conceptual differences
Configuring custom Kerberos principal for Apache Flink
Configuring custom Kerberos principal for SQL Stream Builder
Configuring databases for SSB
Configuring Flink application resources
Configuring Flink applications
Configuring Kafka tables
Configuring Ranger policies for Flink
Configuring Ranger policies for SSB
Configuring RocksDB state backend
Configuring SQL job settings
Connector support in SSB
Connectors
Console Page
Converting DataStreams to Tables
Converting Tables to DataStreams
Core Features of Flink
CREATE Statements
Creating and configuring the HBaseSinkFunction
Creating Atlas entity type definitions for Flink
Creating Flink tables using Templates
Creating Input Transforms
Creating Kafka tables
Creating Kafka tables using Console wizard
Creating Kafka tables using Templates
Creating Materialized Views
Creating new Flink application versions
Creating Tables
Creating User Defined Functions
Creating Webhook tables
CSV format
Data Providers
Data Providers Page
Data Types
Data types for Kafka connector
Datagen connector
DataStream API interoperability
DataStream Applications
DataStream Connectors
Default ports for Flink and SSB
Deployment
Deployment scenarios
Deserialization tab
Developing JavaScript functions
Differences in architecture
Differences in data distribution
Download
Download location of CSA
DROP Statements
Dynamic SQL Hints
Enabling checkpoints for Flink applications
Enabling Flink DEBUG logging
Enabling Kerberos authentication
Enabling Knox authentication
Enabling Knox authentication for Flink Dashboard
Enabling Knox Auto Discovery for Flink
Enabling Knox for SSB in CDP Private Cloud Base 7.1.6
Enabling Knox for SSB in CDP Private Cloud Base 7.1.7
Enabling Knox manually for Flink Dashboard
Enabling savepoints for Flink applications
Enabling security for Apache Flink
Enabling SPNEGO authentication for Flink Dashboard
Enabling TLS for database connection
Encryption in SSB
Enriching streaming data with join
Event Time tab
Example: joining Kafka and Kudu tables
Executing SQL jobs in production mode
Faker connector
Filesystem connector
Fixed issues
Flink API Support
Flink application example
Flink application structure
Flink Dashboard
Flink DDL
Flink DML
Flink metadata collection using Atlas
Flink Project Template
Flink Queries
Flink SQL
Flink SQL and Table API
Flink SQL Overview
Flink Terminology
Functions Tab
Getting Started
Getting Started
Governance
Governance
HBase sink with Flink
History Tab
Hive catalog
Input Transforms
INSERT Statements
Installation
Installing CSD and parcel
Introduction to Materialized Views
Introduction to SQL Stream Builder
JDBC connector
Job Lifecycle
Job Lifecycle
Joining streaming and bounded tables
JSON format
Kafka connector
Kafka connectors
Kafka Metrics Reporter
Kafka with Flink
Key features of SSB
Known issues and limitations
Kudu catalog
Kudu with Flink
Managing connectors and data formats
Managing registered Data Providers
Managing session for SQL jobs
Managing teams in Streaming SQL Console
Managing time in SSB
Materialized Views
Materialized Views Page
Migrating Flink service to a different host
Migrating from Storm to Flink
Migrating SQL jobs
Migration
Monitoring
Monitoring
Monitoring SQL Stream jobs
Other supported statements
Performance & Scalability
Properties tab
Quick Start for Flink
Quick Start for SSB
Reference
Registering Data Providers in SSB
Release Notes
REST API
Resuming Flink applications
Resuming SQL jobs
Running a Flink job
Running a simple Flink application
Running a simple SQL job
Running SQL Stream jobs
Sampling data for a running job
Schema Registry Avro format
Schema Registry catalog
Schema Registry with Flink
Schema tab
Securing Apache Flink
Securing Apache Flink jobs
Securing SQL Stream Builder
Security
Setting parallelism and max parallelism
Setting the Java executable for the Flink client
Setting up MySQL/MariaDB database for SSB
Setting up Oracle database for SSB
Setting up PostgreSQL database for SSB
Setting up your HDFS Home directory
Source, operator and sink in DataStream API
SQL and Table API
SQL and Table API supported features
SQL catalogs for Flink
SQL connectors for Flink
SQL Examples
SQL Jobs Tab
SQL Queries in Flink
SQL Statements in Flink
SQL Stream Applications
SQL Stream Builder architecture
SQL Stream Builder Overview
SSB fails with missing JAR file error
SSB metadata collection using Atlas
Stopping Flink applications
Stopping SQL Stream jobs
Stopping, restarting and editing SQL jobs
Storm Flink Migration
Streaming Analytics in Cloudera
Streaming SQL Console
Streams Messaging Manager integration
Support Matrix
Supported basic data types
Supported data types
System Requirements
Tables Tab
Testing and validating Flink applications
Transformations tab
Unsupported features
Upgrade
Upgrading CSA artifacts for Flink service
Upgrading CSA artifacts for SQL Stream Builder service
Upgrading Flink service
Upgrading SQL Stream Builder service
Uploading and unlocking your keytab in SSB
User Defined Functions
Using Dynamic Materialized View Endpoints
Using EncryptTool for Flink properties
Using Flink CLI
Using SQL Stream Builder REST API
Using SQL Stream Builder with Cloudera Data Visualization
Using System Functions
Using the Streaming SQL Console
Verifying metadata collection
What is Apache Flink?
What's new in Cloudera Streaming Analytics
«
Filter topics
Securing SQL Stream Builder
▶︎
Securing Apache Flink
Authentication and encryption for Flink
▶︎
Enabling security for Apache Flink
Configuring custom Kerberos principal for Apache Flink
Enabling SPNEGO authentication for Flink Dashboard
▶︎
Enabling Knox authentication for Flink Dashboard
Enabling Knox Auto Discovery for Flink
Enabling Knox manually for Flink Dashboard
Accessing the Flink Dashboard through Knox
Configuring Ranger policies for Flink
Securing Apache Flink jobs
Using EncryptTool for Flink properties
▼
Securing SQL Stream Builder
▶︎
Authentication in SSB
▶︎
Enabling Kerberos authentication
Configuring custom Kerberos principal for SQL Stream Builder
▶︎
Enabling Knox authentication
▶︎
Enabling Knox for SSB in CDP Private Cloud Base 7.1.6
Extending Knox with the SSB service definitions
Adding SSB services to the default topologies
Defining the external URLs in SSB
Enabling Knox for SSB in CDP Private Cloud Base 7.1.7
Accessing the Streaming SQL Console through Knox
Uploading and unlocking your keytab in SSB
▶︎
Encryption in SSB
Enabling TLS for database connection
Configuring Ranger policies for SSB
Managing teams in Streaming SQL Console
»
Security
Securing SQL Stream Builder
Authentication in SSB
You can authenticate users to access the Streaming SQL Console using Kerberos or Knox authentication.
Encryption in SSB
When auto-TLS is disabled for the SQL Stream Builder (SSB) service, you must manually set the TLS properties for SSB in Cloudera Manager.
Configuring Ranger policies for SSB
You must add SQL Stream Builder (SSB) service user named ssb to the Ranger policies that are used by Kafka, Schema Registry, Hive and Kudu to provide access to topics, schemas and tables provided by the components.
Managing teams in Streaming SQL Console
You can manage your team, team members and invite new team members under the Teams menu on the Streaming SQL Console.
This site uses cookies and related technologies, as described in our
privacy policy
, for purposes that may include site operation, analytics, enhanced user experience, or advertising. You may choose to consent to our use of these technologies, or
manage your own preferences.
Accept all
1.14.0
1.13.3
1.13.2
1.13.1
1.13.0
1.12.0
1.11.2
1.11.1
1.11.0
1.10.0
1.9.0
1.8.0
1.7.0
1.6
1.6.3
1.6.2
1.6.1
1.6.0
1.5
1.5.3
1.5.1
1.5.0
1.4
1.4.1
1.4.0
1.3.0
1.2.0
1.1.0