KafkaStreamingSourceOptions

Additional options for streaming.

Types

Link copied to clipboard
class Builder
Link copied to clipboard
object Companion

Properties

Link copied to clipboard

When this option is set to 'true', the data output will contain an additional column named "__src_timestamp" that indicates the time when the corresponding record received by the topic. The default value is 'false'. This option is supported in Glue version 4.0 or later.

Link copied to clipboard

The specific TopicPartitions to consume. You must specify at least one of "topicName", "assign" or "subscribePattern".

Link copied to clipboard

A list of bootstrap server URLs, for example, as b-1.vpc-test-2.o4q88o.c6.kafka.us-east-1.amazonaws.com:9094. This option must be specified in the API call or defined in the table metadata in the Data Catalog.

Link copied to clipboard

An optional classification.

Link copied to clipboard

The name of the connection.

Link copied to clipboard

Specifies the delimiter character.

Link copied to clipboard

When this option is set to 'true', for each batch, it will emit the metrics for the duration between the oldest record received by the topic and the time it arrives in Glue to CloudWatch. The metric's name is "glue.driver.streaming.maxConsumerLagInMs". The default value is 'false'. This option is supported in Glue version 4.0 or later.

Link copied to clipboard

The end point when a batch query is ended. Possible values are either "latest" or a JSON string that specifies an ending offset for each TopicPartition.

Link copied to clipboard

Whether to include the Kafka headers. When the option is set to "true", the data output will contain an additional column named "glue_streaming_kafka_headers" with type Array[Struct(key: String, value: String)]. The default value is "false". This option is available in Glue version 3.0 or later only.

Link copied to clipboard

The rate limit on the maximum number of offsets that are processed per trigger interval. The specified total number of offsets is proportionally split across topicPartitions of different volumes. The default value is null, which means that the consumer reads all offsets until the known latest offset.

Link copied to clipboard

The desired minimum number of partitions to read from Kafka. The default value is null, which means that the number of spark partitions is equal to the number of Kafka partitions.

Link copied to clipboard

The number of times to retry before failing to fetch Kafka offsets. The default value is 3.

Link copied to clipboard

The timeout in milliseconds to poll data from Kafka in Spark job executors. The default value is 512.

Link copied to clipboard

The time in milliseconds to wait before retrying to fetch Kafka offsets. The default value is 10.

Link copied to clipboard

The protocol used to communicate with brokers. The possible values are "SSL" or "PLAINTEXT".

Link copied to clipboard

The starting position in the Kafka topic to read data from. The possible values are "earliest" or "latest". The default value is "latest".

Link copied to clipboard

The timestamp of the record in the Kafka topic to start reading data from. The possible values are a timestamp string in UTC format of the pattern yyyy-mm-ddTHH:MM:SSZ (where Z represents a UTC timezone offset with a +/-. For example: "2023-04-04T08:00:00+08:00").

Link copied to clipboard

A Java regex string that identifies the topic list to subscribe to. You must specify at least one of "topicName", "assign" or "subscribePattern".

Link copied to clipboard

The topic name as specified in Apache Kafka. You must specify at least one of "topicName", "assign" or "subscribePattern".

Functions

Link copied to clipboard
Link copied to clipboard
open operator override fun equals(other: Any?): Boolean
Link copied to clipboard
open override fun hashCode(): Int
Link copied to clipboard
open override fun toString(): String