Skip to main content

Event-based

Event-based Sources and Targets

Prophecy supports Kafka Streaming Source and Target. More information on supported Kafka Source and Target options are available here.

The Kafka Gem allows inferring the schema of the events by automatically populating the value column. Schema inference works with both JSON and AVRO file formats. A user is required to provide an example event for schema inference.

Create a Kafka Source Gem

A Kafka Source Gem allows the Streaming Pipeline continuously pull data from a Kafka topic. The following options are supported:

PropertyOptionalDefault ValueComment
Broker ListFalseN/AList of Kafka brokers separated by commas. For eg. kdj-ibg1.us-east-2.aws.cloud:9092, kdj-ibg2.us-east-2.aws.cloud:9092,kdj-ibg3.us-east-2.aws.cloud:9092
Group IDTrueNoneConsumer group ID.
Session TimeoutFalse6000Corresponds to the session.timeout.ms field
Security ProtocolFalseSASL_SSLSupported values are SASL_SSL, PLAINTEXT, SSL, SSL_PLAINTEXT
SASL MechanismsFalseSCRAM-SHA-256SASL mechanism to handle username/password authentication. Supported values are PLAIN, SCRAM-SHA-256 and SCRAM-SHA-512, GSSAPI, OAUTHBEARER
Kafka TopicFalseN/AName of Kafka Topic to Consume

Entering Authentication Credentials

  • Databricks Secrets (recommended): Use Databricks to manage your credentials
  • UserName, Password: Use ONLY for test deployments and during development. This writes credentials to Git repository, which isn't good practice.