Configuring Kafka tables 
 The user-defined Kafka table can be configured based on the schema, event time, input
    transformations and other Kafka specific properties using either the Kafka wizard or
    DDL.
Schema Definition tab  When using the Add Kafka table wizard on the Streaming SQL Console, you can configure   the schema under the Schema tab.Event Time tab  When using the Add Kafka table wizard on the Streaming SQL Console, you can configure   the event time under the Event Time tab.Data Transformations tab  When using the Add Kafka table wizard on the Streaming SQL Console, you can apply input     transformation under the Transformations tab. Input transformations can be used to clean or     arrange the incoming data from the source using javascript functions.Properties tab  When using the Add Kafka table wizard on the Streaming SQL Console, you can configure   the properties under the Properties tab.Deserialization tab  When creating a Kafka table, you can configure how to handle errors due to schema     mismatch using DDL or the Kafka wizard.Assigning Kafka keys in streaming queries  Based on the Sticky Partitioning strategy of Kafka, when null keyed events are sent to a     topic, they are randomly distributed in smaller batches within the partitions.Performance & Scalability  The Kafka and SQL Stream Builder integration enables you to use the Kafka-specific   syntax to customize your SQL queries based on your deployment and use case.