- AbstractConfig - Class in org.apache.kafka.common.config
-
A convenient base class for configurations to extend.
- AbstractConfig(ConfigDef, Map<?, ?>, boolean) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractConfig(ConfigDef, Map<?, ?>) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractConfig(Map<String, Object>) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractProcessor<K,V> - Class in org.apache.kafka.streams.processor
-
- AbstractProcessor() - Constructor for class org.apache.kafka.streams.processor.AbstractProcessor
-
- ACKS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
acks
- addClientSaslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SaslConfigs
-
- addClientSslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SslConfigs
-
- addDeserializerToConfig(Map<String, Object>, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- addDeserializerToConfig(Properties, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- addErrorMessage(String) - Method in class org.apache.kafka.common.config.ConfigValue
-
- addInternalTopic(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds an internal topic
- addLatencySensor(String, String, String, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
- addProcessor(String, ProcessorSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new processor node that receives and processes records output by one or more parent source or processor node.
- addRecord(ConsumerRecord<K, V>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- addSerializerToConfig(Map<String, Object>, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- addSerializerToConfig(Properties, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- addSink(String, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, StreamPartitioner, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic, using
the supplied partitioner.
- addSink(String, String, Serializer, Serializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, Serializer<K>, Serializer<V>, StreamPartitioner<K, V>, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSource(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addSource(String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addStateStore(StateStoreSupplier, boolean, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds a state store
- addStateStore(StateStoreSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds a state store
- advance - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the window's advance interval, i.e.
- advanceBy(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
Returns a window definition with the original size, but advance ("hop") the window by the given
interval, which specifies by how much a window moves forward relative to the previous one.
- after - Variable in class org.apache.kafka.streams.kstream.JoinWindows
-
Maximum time difference for tuples that are after the join tuple.
- after(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are within
the join window interval, and if the timestamp of a record from the secondary stream
is later than or equal to the timestamp of a record from the first stream.
- aggregate(Initializer<T>, Aggregator<K, V, T>, Aggregator<K, V, T>, Serde<T>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate updating values of this stream by the selected key into a new instance of
KTable
.
- aggregate(Initializer<T>, Aggregator<K, V, T>, Aggregator<K, V, T>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate updating values of this stream by the selected key into a new instance of
KTable
using default serializers and deserializers.
- aggregateByKey(Initializer<T>, Aggregator<K, V, T>, Windows<W>, Serde<K>, Serde<T>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Aggregate values of this stream by key on a window basis into a new instance of windowed
KTable
.
- aggregateByKey(Initializer<T>, Aggregator<K, V, T>, Windows<W>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Aggregate values of this stream by key on a window basis into a new instance of windowed
KTable
with default serializers and deserializers.
- aggregateByKey(Initializer<T>, Aggregator<K, V, T>, Serde<K>, Serde<T>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Aggregate values of this stream by key into a new instance of ever-updating
KTable
.
- aggregateByKey(Initializer<T>, Aggregator<K, V, T>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Aggregate values of this stream by key into a new instance of ever-updating
KTable
with default serializers and deserializers.
- Aggregator<K,V,T> - Interface in org.apache.kafka.streams.kstream
-
The
Aggregator
interface for aggregating values of the given key.
- all() - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Return an iterator over all keys in the database.
- allPartitionsSorted(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- AlreadyExistsException - Exception in org.apache.kafka.connect.errors
-
Indicates the operation tried to create an entity that already exists.
- AlreadyExistsException(String) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- AlreadyExistsException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- AlreadyExistsException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- ApiException - Exception in org.apache.kafka.common.errors
-
Any API exception that is part of the public protocol and should be a subclass of this class and be part of this
package.
- ApiException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException(String) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException(Throwable) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException() - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- APPLICATION_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
application.id
- APPLICATION_ID_DOC - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- applicationId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the application id
- apply(K, V, T) - Method in interface org.apache.kafka.streams.kstream.Aggregator
-
Compute a new aggregate from the key and value of a record and the current aggregate of the same key.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.ForeachAction
-
Perform an action for each record of a stream.
- apply() - Method in interface org.apache.kafka.streams.kstream.Initializer
-
Return the initial value for an aggregation.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.KeyValueMapper
-
Map a record with the given key and value to a new value.
- apply(V, V) - Method in interface org.apache.kafka.streams.kstream.Reducer
-
Aggregate the two given values into a single one.
- apply(V1, V2) - Method in interface org.apache.kafka.streams.kstream.ValueJoiner
-
Return a joined value consisting of value1
and value2
.
- apply(V1) - Method in interface org.apache.kafka.streams.kstream.ValueMapper
-
Map the given value to a new value.
- array(Schema) - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- assign(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- assign(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Manually assign a list of partition to this consumer.
- assign(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- assign(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RangeAssignor
-
- assign(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- assignment() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- assignment() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the set of partitions currently assigned to this consumer.
- assignment() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- assignment() - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Get the current set of assigned TopicPartitions for this task.
- atLeast(Number) - Static method in class org.apache.kafka.common.config.ConfigDef.Range
-
A numeric range that checks only the lower bound
- AuthenticationException - Exception in org.apache.kafka.common.errors
-
- AuthenticationException(String) - Constructor for exception org.apache.kafka.common.errors.AuthenticationException
-
- AuthenticationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.AuthenticationException
-
- AuthorizationException - Exception in org.apache.kafka.common.errors
-
- AuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.AuthorizationException
-
- AuthorizationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.AuthorizationException
-
- AUTO_COMMIT_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
auto.commit.interval.ms
- AUTO_OFFSET_RESET_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
auto.offset.reset
- AUTO_OFFSET_RESET_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- availablePartitionsForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the list of available partitions for this topic
- BATCH_SIZE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
batch.size
- before - Variable in class org.apache.kafka.streams.kstream.JoinWindows
-
Maximum time difference for tuples that are before the join tuple.
- before(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are within
the join window interval, and if the timestamp of a record from the secondary stream is
earlier than or equal to the timestamp of a record from the first stream.
- between(Number, Number) - Static method in class org.apache.kafka.common.config.ConfigDef.Range
-
A numeric range that checks both the upper and lower bound
- BLOCK_ON_BUFFER_FULL_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- bool() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- BOOLEAN_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- bootstrap(List<InetSocketAddress>) - Static method in class org.apache.kafka.common.Cluster
-
Create a "bootstrap" cluster using the given list of host/ports
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
bootstrap.servers
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
bootstrap.servers
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
bootstrap.servers
- branch(Predicate<K, V>...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Creates an array of
KStream
from this stream by branching the elements in the original stream based on the supplied predicates.
- BrokerNotAvailableException - Exception in org.apache.kafka.common.errors
-
- BrokerNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.BrokerNotAvailableException
-
- BrokerNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.BrokerNotAvailableException
-
- BUFFER_MEMORY_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
buffer.memory
- BUFFERED_RECORDS_PER_PARTITION_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
buffered.records.per.partition
- BufferExhaustedException - Exception in org.apache.kafka.clients.producer
-
This exception is thrown if the producer is in non-blocking mode and the rate of data production exceeds the rate at
which data can be sent for long enough for the alloted buffer to be exhausted.
- BufferExhaustedException(String) - Constructor for exception org.apache.kafka.clients.producer.BufferExhaustedException
-
- build() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Build the Schema using the current settings
- build(String, Integer) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Build the topology for the specified topic group.
- build() - Method in interface org.apache.kafka.streams.state.Stores.InMemoryKeyValueFactory
-
Return the instance of StateStoreSupplier of new key-value store.
- build() - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Return the instance of StateStoreSupplier of new key-value store.
- builder() - Static method in class org.apache.kafka.connect.data.Date
-
Returns a SchemaBuilder for a Date.
- builder(int) - Static method in class org.apache.kafka.connect.data.Decimal
-
Returns a SchemaBuilder for a Decimal with the given scale factor.
- builder() - Static method in class org.apache.kafka.connect.data.Time
-
Returns a SchemaBuilder for a Time.
- builder() - Static method in class org.apache.kafka.connect.data.Timestamp
-
Returns a SchemaBuilder for a Timestamp.
- ByteArray() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- ByteArrayDeserializer - Class in org.apache.kafka.common.serialization
-
- ByteArrayDeserializer() - Constructor for class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- ByteArraySerializer - Class in org.apache.kafka.common.serialization
-
- ByteArraySerializer() - Constructor for class org.apache.kafka.common.serialization.ByteArraySerializer
-
- ByteBuffer() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- ByteBufferDeserializer - Class in org.apache.kafka.common.serialization
-
- ByteBufferDeserializer() - Constructor for class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- ByteBufferSerializer - Class in org.apache.kafka.common.serialization
-
- ByteBufferSerializer() - Constructor for class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- Bytes() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- bytes() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- BYTES_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- BytesDeserializer - Class in org.apache.kafka.common.serialization
-
- BytesDeserializer() - Constructor for class org.apache.kafka.common.serialization.BytesDeserializer
-
- BytesSerializer - Class in org.apache.kafka.common.serialization
-
- BytesSerializer() - Constructor for class org.apache.kafka.common.serialization.BytesSerializer
-
- Callback - Interface in org.apache.kafka.clients.producer
-
A callback interface that the user can implement to allow code to execute when the request is complete.
- CHECK_CRCS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
check.crcs
- checksum() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The checksum (CRC32) of the record.
- checksum() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The checksum (CRC32) of the record.
- cleanUp() - Method in class org.apache.kafka.streams.KafkaStreams
-
Cleans up local state store directory (state.dir
), by deleting all data with regard to the application-id.
- clear() - Method in class org.apache.kafka.clients.producer.MockProducer
-
Clear the stored history of sent records
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
client.id
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
client.id
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
client.id
- close() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- close() - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
This is called when interceptor is closed
- close() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Close the consumer, waiting indefinitely for any needed cleanup.
- close() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- close() - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Close this producer.
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
This method waits up to timeout
for the producer to complete the sending of all incomplete requests.
- close() - Method in class org.apache.kafka.clients.producer.MockProducer
-
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.producer.MockProducer
-
- close() - Method in interface org.apache.kafka.clients.producer.Partitioner
-
This is called when partitioner is closed.
- close() - Method in interface org.apache.kafka.clients.producer.Producer
-
Close this producer
- close(long, TimeUnit) - Method in interface org.apache.kafka.clients.producer.Producer
-
Tries to close the producer cleanly within the specified timeout.
- close() - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
This is called when interceptor is closed
- close() - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- close() - Method in interface org.apache.kafka.common.serialization.Deserializer
-
- close() - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- close() - Method in interface org.apache.kafka.common.serialization.Serde
-
Close this serde class, which will close the underlying serializer and deserializer.
- close() - Method in interface org.apache.kafka.common.serialization.Serializer
-
Close this serializer.
- close() - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- close(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
The SinkTask use this method to close writers for partitions that are no
longer assigned to the SinkTask.
- close() - Method in class org.apache.kafka.streams.KafkaStreams
-
Shutdown this stream instance by signaling all the threads to stop,
and then wait for them to join.
- close() - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Close this processor and clean up any resources.
- close() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.Processor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Close the storage engine.
- close() - Method in interface org.apache.kafka.streams.state.KeyValueIterator
-
- close() - Method in interface org.apache.kafka.streams.state.WindowStoreIterator
-
- closed() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- Cluster - Class in org.apache.kafka.common
-
A representation of a subset of the nodes, topics, and partitions in the Kafka cluster.
- Cluster(Collection<Node>, Collection<PartitionInfo>, Set<String>) - Constructor for class org.apache.kafka.common.Cluster
-
Create a new cluster with the given nodes and partitions
- ClusterAuthorizationException - Exception in org.apache.kafka.common.errors
-
- ClusterAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.ClusterAuthorizationException
-
- ClusterAuthorizationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ClusterAuthorizationException
-
- commit() - Method in class org.apache.kafka.connect.source.SourceTask
-
- commit() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Requests a commit
- COMMIT_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
commit.interval.ms
- commitAsync() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync(OffsetCommitCallback) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for all the subscribed list of topics and partition.
- commitAsync(OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for the subscribed list of topics and partitions.
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit the specified offsets for the specified list of topics and partitions to Kafka.
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitAsync() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitAsync(OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- CommitFailedException - Exception in org.apache.kafka.clients.consumer
-
- CommitFailedException(String) - Constructor for exception org.apache.kafka.clients.consumer.CommitFailedException
-
- commitRecord(SourceRecord) - Method in class org.apache.kafka.connect.source.SourceTask
-
Commit an individual
SourceRecord
when the callback from the producer client is received.
- commitSync() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitSync() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for all the subscribed list of topics and partitions.
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit the specified offsets for the specified list of topics and partitions.
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitSync() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- committed(TopicPartition) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- committed(TopicPartition) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the last committed offset for the given partition (whether the commit happened by this process or
another).
- committed(TopicPartition) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- compareTo(TaskId) - Method in class org.apache.kafka.streams.processor.TaskId
-
- completeNext() - Method in class org.apache.kafka.clients.producer.MockProducer
-
Complete the earliest uncompleted call successfully.
- COMPRESSION_TYPE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
compression.type
- Config - Class in org.apache.kafka.common.config
-
- Config(List<ConfigValue>) - Constructor for class org.apache.kafka.common.config.Config
-
- config() - Method in class org.apache.kafka.connect.connector.Connector
-
Define the configuration for the connector.
- ConfigDef - Class in org.apache.kafka.common.config
-
This class is used for specifying the set of expected configurations.
- ConfigDef() - Constructor for class org.apache.kafka.common.config.ConfigDef
-
- ConfigDef.ConfigKey - Class in org.apache.kafka.common.config
-
- ConfigDef.ConfigKey(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Constructor for class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- ConfigDef.Importance - Enum in org.apache.kafka.common.config
-
The importance level for a configuration
- ConfigDef.Range - Class in org.apache.kafka.common.config
-
Validation logic for numeric ranges
- ConfigDef.Recommender - Interface in org.apache.kafka.common.config
-
This is used by the
ConfigDef.validate(Map)
to get valid values for a configuration given the current
configuration values in order to perform full configuration validation and visibility modification.
- ConfigDef.Type - Enum in org.apache.kafka.common.config
-
The config types
- ConfigDef.Validator - Interface in org.apache.kafka.common.config
-
Validation logic the user may provide to perform single configuration validation.
- ConfigDef.ValidString - Class in org.apache.kafka.common.config
-
- ConfigDef.Width - Enum in org.apache.kafka.common.config
-
The width of a configuration value
- ConfigException - Exception in org.apache.kafka.common.config
-
Thrown if the user supplies an invalid configuration
- ConfigException(String) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- ConfigException(String, Object) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- ConfigException(String, Object, String) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- configKeys() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the configuration keys
- Configurable - Interface in org.apache.kafka.common
-
A Mix-in style interface for classes that are instantiated by reflection and need to take configuration parameters
- configure(Map<String, ?>) - Method in interface org.apache.kafka.common.Configurable
-
Configure this class with the given key-value pairs
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Deserializer
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Serde
-
Configure this class, which will configure the underlying serializer and deserializer.
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Serializer
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.connect.storage.Converter
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- ConfigValue - Class in org.apache.kafka.common.config
-
- ConfigValue(String) - Constructor for class org.apache.kafka.common.config.ConfigValue
-
- ConfigValue(String, Object, List<Object>, List<String>) - Constructor for class org.apache.kafka.common.config.ConfigValue
-
- configValues() - Method in class org.apache.kafka.common.config.Config
-
- ConnectException - Exception in org.apache.kafka.connect.errors
-
ConnectException is the top-level exception type generated by Kafka Connect and connector implementations.
- ConnectException(String) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- ConnectException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- ConnectException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
connections.max.idle.ms
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
connections.max.idle.ms
- Connector - Class in org.apache.kafka.connect.connector
-
Connectors manage integration of Kafka Connect with another system, either as an input that ingests
data into Kafka or an output that passes data to an external system.
- Connector() - Constructor for class org.apache.kafka.connect.connector.Connector
-
- ConnectorContext - Interface in org.apache.kafka.connect.connector
-
ConnectorContext allows Connectors to proactively interact with the Kafka Connect runtime.
- ConnectorUtils - Class in org.apache.kafka.connect.util
-
Utilities that connector implementations might find useful.
- ConnectorUtils() - Constructor for class org.apache.kafka.connect.util.ConnectorUtils
-
- connectProcessorAndStateStores(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects the processor and the state stores
- connectProcessors(String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects a list of processors.
- ConnectRecord - Class in org.apache.kafka.connect.connector
-
Base class for records containing data to be copied to/from Kafka.
- ConnectRecord(String, Integer, Schema, Object) - Constructor for class org.apache.kafka.connect.connector.ConnectRecord
-
- ConnectRecord(String, Integer, Schema, Object, Schema, Object) - Constructor for class org.apache.kafka.connect.connector.ConnectRecord
-
- ConnectSchema - Class in org.apache.kafka.connect.data
-
- ConnectSchema(Schema.Type, boolean, Object, String, Integer, String, Map<String, String>, List<Field>, Schema, Schema) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a Schema.
- ConnectSchema(Schema.Type, boolean, Object, String, Integer, String) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a Schema for a primitive type, setting schema parameters, struct fields, and key and value schemas to null.
- ConnectSchema(Schema.Type) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a default schema for a primitive type.
- Consumer<K,V> - Interface in org.apache.kafka.clients.consumer
-
- ConsumerConfig - Class in org.apache.kafka.clients.consumer
-
The consumer configuration keys
- ConsumerInterceptor<K,V> - Interface in org.apache.kafka.clients.consumer
-
A plugin interface that allows you to intercept (and possibly mutate) records received by the consumer.
- ConsumerRebalanceListener - Interface in org.apache.kafka.clients.consumer
-
A callback interface that the user can implement to trigger custom actions when the set of partitions assigned to the
consumer changes.
- ConsumerRecord<K,V> - Class in org.apache.kafka.clients.consumer
-
A key/value pair to be received from Kafka.
- ConsumerRecord(String, int, long, K, V) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecord
-
Creates a record to be received from a specified topic and partition (provided for
compatibility with Kafka 0.9 before the message format supported timestamps and before
serialized metadata were exposed).
- ConsumerRecord(String, int, long, long, TimestampType, long, int, int, K, V) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecord
-
Creates a record to be received from a specified topic and partition
- ConsumerRecords<K,V> - Class in org.apache.kafka.clients.consumer
-
A container that holds the list
ConsumerRecord
per partition for a
particular topic.
- ConsumerRecords(Map<TopicPartition, List<ConsumerRecord<K, V>>>) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecords
-
- ConsumerRecordTimestampExtractor - Class in org.apache.kafka.streams.processor
-
Retrieves built-in timestamps from Kafka messages (introduced in KIP-32: Add timestamps to Kafka message).
- ConsumerRecordTimestampExtractor() - Constructor for class org.apache.kafka.streams.processor.ConsumerRecordTimestampExtractor
-
- context - Variable in class org.apache.kafka.connect.connector.Connector
-
- context - Variable in class org.apache.kafka.connect.sink.SinkTask
-
- context - Variable in class org.apache.kafka.connect.source.SourceTask
-
- context() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- ControllerMovedException - Exception in org.apache.kafka.common.errors
-
- ControllerMovedException(String) - Constructor for exception org.apache.kafka.common.errors.ControllerMovedException
-
- ControllerMovedException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ControllerMovedException
-
- Converter - Interface in org.apache.kafka.connect.storage
-
The Converter interface provides support for translating between Kafka Connect's runtime data format
and byte[].
- convertToString(Object, ConfigDef.Type) - Static method in class org.apache.kafka.common.config.ConfigDef
-
- copartitionGroups() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the copartition groups.
- copartitionSources(Collection<String>) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Asserts that the streams of the specified source nodes must be copartitioned.
- CorruptRecordException - Exception in org.apache.kafka.common.errors
-
This exception indicates a record has failed its internal CRC check, this generally indicates network or disk
corruption.
- CorruptRecordException() - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(String) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(Throwable) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- count() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
The number of records for all topics
- count(String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of this stream by the selected key into a new instance of
KTable
.
- countByKey(Windows<W>, Serde<K>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Count number of records of this stream by key on a window basis into a new instance of windowed
KTable
.
- countByKey(Windows<W>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Count number of records of this stream by key on a window basis into a new instance of windowed
KTable
with default serializers and deserializers.
- countByKey(Serde<K>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Count number of records of this stream by key into a new instance of ever-updating
KTable
.
- countByKey(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Count number of records of this stream by key into a new instance of ever-updating
KTable
with default serializers and deserializers.
- create(String) - Static method in class org.apache.kafka.streams.state.Stores
-
- DataException - Exception in org.apache.kafka.connect.errors
-
Base class for all Kafka Connect data API exceptions.
- DataException(String) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- DataException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- DataException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- Date - Class in org.apache.kafka.connect.data
-
A date representing a calendar day with no time of day or timezone.
- Date() - Constructor for class org.apache.kafka.connect.data.Date
-
- Decimal - Class in org.apache.kafka.connect.data
-
An arbitrary-precision signed decimal number.
- Decimal() - Constructor for class org.apache.kafka.connect.data.Decimal
-
- DEFAULT_EXCLUDE_INTERNAL_TOPICS - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- DEFAULT_KERBEROS_KINIT_CMD - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_MIN_TIME_BEFORE_RELOGIN - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_TICKET_RENEW_JITTER - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_TICKET_RENEW_WINDOW_FACTOR - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_MAX_PARTITION_FETCH_BYTES - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- DEFAULT_PRINCIPAL_BUILDER_CLASS - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SASL_ENABLED_MECHANISMS - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SASL_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SSL_ENABLED_PROTOCOLS - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_KEYMANGER_ALGORITHM - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_KEYSTORE_TYPE - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_PROTOCOL - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_TRUSTMANAGER_ALGORITHM - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_TRUSTSTORE_TYPE - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DefaultPartitionGrouper - Class in org.apache.kafka.streams.processor
-
Default implementation of the
PartitionGrouper
interface that groups partitions by the partition id.
- DefaultPartitionGrouper() - Constructor for class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
- defaultValue - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- defaultValue() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- defaultValue() - Method in interface org.apache.kafka.connect.data.Schema
-
- defaultValue() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- defaultValue(Object) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the default value for this schema.
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no dependents
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic, not dependents and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value and no special validation logic
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic, no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no group, no order in group, no width, no display name, no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic
- define(String, ConfigDef.Type, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value and no special validation logic
- delete(K) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Delete the value from the store (if there is one)
- dependents - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- description() - Method in class org.apache.kafka.common.MetricName
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- deserialize(String, byte[]) - Method in interface org.apache.kafka.common.serialization.Deserializer
-
Deserialize a record value from a bytearray into a value or object.
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- Deserializer<T> - Interface in org.apache.kafka.common.serialization
-
- deserializer() - Method in interface org.apache.kafka.common.serialization.Serde
-
- DisconnectException - Exception in org.apache.kafka.common.errors
-
Server disconnected before a request could be completed.
- DisconnectException() - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(String) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(Throwable) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- displayName - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- doc() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- doc() - Method in interface org.apache.kafka.connect.data.Schema
-
- doc() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- doc(String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the documentation for this schema.
- documentation - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- Double() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- DoubleDeserializer - Class in org.apache.kafka.common.serialization
-
- DoubleDeserializer() - Constructor for class org.apache.kafka.common.serialization.DoubleDeserializer
-
- DoubleSerializer - Class in org.apache.kafka.common.serialization
-
- DoubleSerializer() - Constructor for class org.apache.kafka.common.serialization.DoubleSerializer
-
- fetch(K, long, long) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Get all the key-value pairs with the given key and the time range from all
the existing windows.
- FETCH_MAX_WAIT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
fetch.max.wait.ms
- FETCH_MIN_BYTES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
fetch.min.bytes
- field(String) - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- Field - Class in org.apache.kafka.connect.data
-
A field in a
Struct
, consisting of a field name, index, and
Schema
for the field value.
- Field(String, int, Schema) - Constructor for class org.apache.kafka.connect.data.Field
-
- field(String) - Method in interface org.apache.kafka.connect.data.Schema
-
Get a field for this Schema by name.
- field(String, Schema) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Add a field to this struct schema.
- field(String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- fields() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- fields() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the list of fields for this Schema.
- fields() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Get the list of fields for this Schema.
- fillInStackTrace() - Method in exception org.apache.kafka.common.errors.ApiException
-
- fillInStackTrace() - Method in exception org.apache.kafka.common.errors.SerializationException
-
- filter(Predicate<K, V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new instance of
KStream
that consists of all elements of this stream which satisfy a predicate.
- filter(Predicate<K, V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new instance of
KTable
that consists of all elements of this stream which satisfy a predicate.
- filterNot(Predicate<K, V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new instance of
KStream
that consists all elements of this stream which do not satisfy a predicate.
- filterNot(Predicate<K, V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new instance of
KTable
that consists all elements of this stream which do not satisfy a predicate.
- flatMap(KeyValueMapper<K, V, Iterable<KeyValue<K1, V1>>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new instance of
KStream
by transforming each element in this stream into zero or more elements in the new stream.
- flatMapValues(ValueMapper<V, Iterable<V1>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new instance of
KStream
by transforming the value of each element in this stream into zero or more values with the same key in the new stream.
- float32() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- FLOAT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- float64() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- FLOAT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- flush() - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Invoking this method makes all buffered records immediately available to send (even if linger.ms
is
greater than 0) and blocks on the completion of the requests associated with these records.
- flush() - Method in class org.apache.kafka.clients.producer.MockProducer
-
- flush() - Method in interface org.apache.kafka.clients.producer.Producer
-
Flush any accumulated records from the producer.
- flush(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- flush() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Flush any cached data
- foreach(ForeachAction<K, V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Perform an action on each element of
KStream
.
- foreach(ForeachAction<K, V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Perform an action on each element of
KTable
.
- ForeachAction<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
ForeachAction
interface for performing an action on a key-value pair.
- forward(K, V) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to the downstream processors
- forward(K, V, int) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by childIndex
- forward(K, V, String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by the downstream processor name
- fromConnectData(String, Schema, Object) - Method in interface org.apache.kafka.connect.storage.Converter
-
Convert a Kafka Connect data object to a native object for serialization.
- fromConnectData(String, Schema, Object) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Date
-
Convert a value from its logical format (Date) to it's encoded format.
- fromLogical(Schema, BigDecimal) - Static method in class org.apache.kafka.connect.data.Decimal
-
Convert a value from its logical format (BigDecimal) to it's encoded format.
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Time
-
Convert a value from its logical format (Time) to it's encoded format.
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Timestamp
-
Convert a value from its logical format (Date) to it's encoded format.
- get(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- get(String) - Method in class org.apache.kafka.connect.data.Struct
-
Get the value of a field, returning the default value if no value has been set yet and a default value is specified
in the field's schema.
- get(Field) - Method in class org.apache.kafka.connect.data.Struct
-
Get the value of a field, returning the default value if no value has been set yet and a default value is specified
in the field's schema.
- get() - Method in interface org.apache.kafka.streams.kstream.TransformerSupplier
-
- get() - Method in interface org.apache.kafka.streams.kstream.ValueTransformerSupplier
-
- get() - Method in interface org.apache.kafka.streams.processor.ProcessorSupplier
-
- get() - Method in interface org.apache.kafka.streams.processor.StateStoreSupplier
-
- get(K) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Get the value corresponding to this key
- getArray(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getBoolean(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getBoolean(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getBytes(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getClass(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getConfiguredInstance(String, Class<T>) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get a configured instance of the give class specified by the given configuration key.
- getConfiguredInstances(String, Class<T>) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get a list of configured instances of the given class specified by the given configuration key.
- getConsumer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
Creates an instance of Consumer which is used to consume records of source topics.
- getConsumerConfigs(StreamThread, String, String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getDouble(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getFloat32(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getFloat64(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getInt16(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt32(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt64(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt8(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getList(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getLong(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getMap(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getName() - Method in enum org.apache.kafka.connect.data.Schema.Type
-
- getPassword(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getProducer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
Creates an instance of Producer which is used to produce records.
- getProducerConfigs(String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getRestoreConsumer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
Creates an instance of Consumer which is used to consume records of internal topics.
- getRestoreConsumerConfigs(String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getShort(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getStateStore(String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Get the state store given the store name.
- getString(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getString(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getStruct(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getWithoutDefault(String) - Method in class org.apache.kafka.connect.data.Struct
-
Get the underlying raw value for the field without accounting for default values.
- group - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- group() - Method in class org.apache.kafka.common.MetricName
-
- GROUP_ID_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
group.id
- GroupAuthorizationException - Exception in org.apache.kafka.common.errors
-
- GroupAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.GroupAuthorizationException
-
- groupBy(KeyValueMapper<K, V, KeyValue<K1, V1>>, Serde<K1>, Serde<V1>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- groupBy(KeyValueMapper<K, V, KeyValue<K1, V1>>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Group the records of this
KTable
using the provided
KeyValueMapper
and default serializers and deserializers.
- GroupCoordinatorNotAvailableException - Exception in org.apache.kafka.common.errors
-
The broker returns this error code for consumer metadata requests or offset commit requests if the offsets topic has
not yet been created.
- GroupCoordinatorNotAvailableException() - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- groupId() - Method in exception org.apache.kafka.common.errors.GroupAuthorizationException
-
- GroupLoadInProgressException - Exception in org.apache.kafka.common.errors
-
The broker returns this error code for any coordinator request if it is still loading the metadata (after a leader change
for that offsets topic partition) for this group.
- GroupLoadInProgressException() - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(String) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- groupPartitions(List<T>, int) - Static method in class org.apache.kafka.connect.util.ConnectorUtils
-
Given a list of elements and a target number of groups, generates list of groups of
elements to match the target number of groups, spreading them evenly among the groups.
- groups() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the groups for the configuration
- GSSAPI_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- id() - Method in class org.apache.kafka.common.Node
-
The node id of this node
- idString() - Method in class org.apache.kafka.common.Node
-
String representation of the node id.
- ignore(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- IllegalGenerationException - Exception in org.apache.kafka.common.errors
-
- IllegalGenerationException() - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(String) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalSaslStateException - Exception in org.apache.kafka.common.errors
-
- IllegalSaslStateException(String) - Constructor for exception org.apache.kafka.common.errors.IllegalSaslStateException
-
- IllegalSaslStateException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalSaslStateException
-
- IllegalWorkerStateException - Exception in org.apache.kafka.connect.errors
-
Indicates that a method has been invoked illegally or at an invalid time by a connector or task.
- IllegalWorkerStateException(String) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- IllegalWorkerStateException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- IllegalWorkerStateException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- importance - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- in(String...) - Static method in class org.apache.kafka.common.config.ConfigDef.ValidString
-
- InconsistentGroupProtocolException - Exception in org.apache.kafka.common.errors
-
- InconsistentGroupProtocolException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InconsistentGroupProtocolException
-
- InconsistentGroupProtocolException(String) - Constructor for exception org.apache.kafka.common.errors.InconsistentGroupProtocolException
-
- index() - Method in class org.apache.kafka.connect.data.Field
-
Get the index of this field within the struct.
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Initialize this transformer with the given context.
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Initialize this transformer with the given context.
- init(ProcessorContext) - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.processor.Processor
-
Initialize this processor with the given context.
- init(ProcessorContext, StateStore) - Method in interface org.apache.kafka.streams.processor.StateStore
-
Initializes this state store
- initialize(ConnectorContext) - Method in class org.apache.kafka.connect.connector.Connector
-
Initialize this connector, using the provided ConnectorContext to notify the runtime of
input configuration changes.
- initialize(ConnectorContext, List<Map<String, String>>) - Method in class org.apache.kafka.connect.connector.Connector
-
Initialize this connector, using the provided ConnectorContext to notify the runtime of
input configuration changes and using the provided set of Task configurations.
- initialize(SinkTaskContext) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Initialize the context of this task.
- initialize(SourceTaskContext) - Method in class org.apache.kafka.connect.source.SourceTask
-
Initialize this SourceTask with the specified context object.
- Initializer<T> - Interface in org.apache.kafka.streams.kstream
-
The
Initializer
interface for creating an initial value in aggregations.
- inMemory() - Method in interface org.apache.kafka.streams.state.Stores.KeyValueFactory
-
Keep all key-value entries in-memory, although for durability all entries are recorded in a Kafka topic that can be
read to restore the entries if they are lost.
- INSTANCE - Static variable in exception org.apache.kafka.common.errors.DisconnectException
-
- INSTANCE - Static variable in exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- inSyncReplicas() - Method in class org.apache.kafka.common.PartitionInfo
-
The subset of the replicas that are in sync, that is caught-up to the leader and ready to take over as leader if
the leader should fail
- int16() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT16_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int32() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int64() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int8() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT8_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- Integer() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- IntegerDeserializer - Class in org.apache.kafka.common.serialization
-
- IntegerDeserializer() - Constructor for class org.apache.kafka.common.serialization.IntegerDeserializer
-
- IntegerSerializer - Class in org.apache.kafka.common.serialization
-
- IntegerSerializer() - Constructor for class org.apache.kafka.common.serialization.IntegerSerializer
-
- INTERCEPTOR_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
interceptor.classes
- INTERCEPTOR_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
interceptor.classes
- INTERCEPTOR_CLASSES_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- INTERCEPTOR_CLASSES_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- InterruptException - Exception in org.apache.kafka.common.errors
-
An unchecked wrapper for InterruptedException
- InterruptException(InterruptedException) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- InterruptException(String, InterruptedException) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- InterruptException(String) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- interSourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- InvalidCommitOffsetSizeException - Exception in org.apache.kafka.common.errors
-
- InvalidCommitOffsetSizeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidCommitOffsetSizeException
-
- InvalidCommitOffsetSizeException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidCommitOffsetSizeException
-
- InvalidFetchSizeException - Exception in org.apache.kafka.common.errors
-
- InvalidFetchSizeException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidFetchSizeException
-
- InvalidFetchSizeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidFetchSizeException
-
- InvalidGroupIdException - Exception in org.apache.kafka.common.errors
-
- InvalidGroupIdException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidGroupIdException
-
- InvalidGroupIdException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidGroupIdException
-
- InvalidMetadataException - Exception in org.apache.kafka.common.errors
-
An exception that may indicate the client's metadata is out of date
- InvalidMetadataException() - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidOffsetException - Exception in org.apache.kafka.clients.consumer
-
Thrown when the offset for a set of partitions is invalid (either undefined or out of range),
and no reset policy has been configured.
- InvalidOffsetException(String) - Constructor for exception org.apache.kafka.clients.consumer.InvalidOffsetException
-
- InvalidOffsetException - Exception in org.apache.kafka.common.errors
-
Thrown when the offset for a set of partitions is invalid (either undefined or out of range),
and no reset policy has been configured.
- InvalidOffsetException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidOffsetException
-
- InvalidOffsetException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidOffsetException
-
- InvalidRequiredAcksException - Exception in org.apache.kafka.common.errors
-
- InvalidRequiredAcksException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidRequiredAcksException
-
- InvalidSessionTimeoutException - Exception in org.apache.kafka.common.errors
-
- InvalidSessionTimeoutException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidSessionTimeoutException
-
- InvalidSessionTimeoutException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidSessionTimeoutException
-
- InvalidTimestampException - Exception in org.apache.kafka.common.errors
-
Indicate the timestamp of a record is invalid.
- InvalidTimestampException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidTimestampException
-
- InvalidTimestampException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTimestampException
-
- InvalidTopicException - Exception in org.apache.kafka.common.errors
-
The client has attempted to perform an operation on an invalid topic.
- InvalidTopicException() - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- isBootstrapConfigured() - Method in class org.apache.kafka.common.Cluster
-
- isEmpty() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
- isEmpty() - Method in class org.apache.kafka.common.Node
-
Check whether this node is empty, which may be the case if noNode() is used as a placeholder
in a response payload with an error.
- isOptional() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- isOptional() - Method in interface org.apache.kafka.connect.data.Schema
-
- isOptional() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- isPrimitive() - Method in enum org.apache.kafka.connect.data.Schema.Type
-
- iterator() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
- KafkaClientSupplier - Interface in org.apache.kafka.streams
-
- KafkaConsumer<K,V> - Class in org.apache.kafka.clients.consumer
-
A Kafka client that consumes records from a Kafka cluster.
- KafkaConsumer(Map<String, Object>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a set of key-value pairs as configuration.
- KafkaConsumer(Map<String, Object>, Deserializer<K>, Deserializer<V>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a set of key-value pairs as configuration, and a key and a value
Deserializer
.
- KafkaConsumer(Properties) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a
Properties
object as configuration.
- KafkaConsumer(Properties, Deserializer<K>, Deserializer<V>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a
Properties
object as configuration, and a
key and a value
Deserializer
.
- KafkaException - Exception in org.apache.kafka.common
-
The base class of all other Kafka exceptions
- KafkaException(String, Throwable) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException(String) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException(Throwable) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException() - Constructor for exception org.apache.kafka.common.KafkaException
-
- kafkaOffset() - Method in class org.apache.kafka.connect.sink.SinkRecord
-
- kafkaPartition() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- KafkaProducer<K,V> - Class in org.apache.kafka.clients.producer
-
A Kafka client that publishes records to the Kafka cluster.
- KafkaProducer(Map<String, Object>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration.
- KafkaProducer(Map<String, Object>, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration, a key and a value
Serializer
.
- KafkaProducer(Properties) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration.
- KafkaProducer(Properties, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration, a key and a value
Serializer
.
- KafkaStreams - Class in org.apache.kafka.streams
-
Kafka Streams allows for performing continuous computation on input coming from one or more input topics and
sends output to zero or more output topics.
- KafkaStreams(TopologyBuilder, Properties) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Construct the stream instance.
- KafkaStreams(TopologyBuilder, StreamsConfig) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Construct the stream instance.
- KafkaStreams(TopologyBuilder, StreamsConfig, KafkaClientSupplier) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Construct the stream instance.
- key() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The key (or null if no key is specified)
- key() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- key() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- key - Variable in class org.apache.kafka.streams.KeyValue
-
The key of the key-value pair.
- key() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the key of the window.
- KEY_DESERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
key.deserializer
- KEY_DESERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- KEY_SERDE_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
key.serde
- KEY_SERDE_CLASS_DOC - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- KEY_SERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
key.serializer
- KEY_SERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- keyDeserializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key deserializer.
- keyFrom(byte[]) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Deserialize the key from raw bytes.
- keySchema() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- keySchema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- keySchema() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the key schema for this map schema.
- keySchema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- keySerde() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the default key serde
- keySerde() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serde.
- keySerde() - Method in class org.apache.kafka.streams.StreamsConfig
-
- keySerializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serializer.
- KeyValue<K,V> - Class in org.apache.kafka.streams
-
A key-value pair defined for a single Kafka Streams record.
- KeyValue(K, V) - Constructor for class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- KeyValueIterator<K,V> - Interface in org.apache.kafka.streams.state
-
- KeyValueMapper<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
The
KeyValueMapper
interface for mapping a key-value pair to a new value (could be another key-value pair).
- KeyValueStore<K,V> - Interface in org.apache.kafka.streams.state
-
A key-value store that supports put/get/delete and range queries.
- KGroupedTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KGroupedTable
is an abstraction of a
grouped changelog stream from a primary-keyed table,
usually on a different grouping key than the original primary key.
- KStream<K,V> - Interface in org.apache.kafka.streams.kstream
-
KStream
is an abstraction of a
record stream of key-value pairs.
- KStreamBuilder - Class in org.apache.kafka.streams.kstream
-
KStreamBuilder
is a subclass of
TopologyBuilder
that provides the Kafka Streams DSL
for users to specify computational logic and translates the given logic to a
org.apache.kafka.streams.processor.internals.ProcessorTopology
.
- KStreamBuilder() - Constructor for class org.apache.kafka.streams.kstream.KStreamBuilder
-
- KTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KTable
is an abstraction of a
changelog stream from a primary-keyed table.
- of(String) - Static method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are equal.
- of(String, long) - Static method in class org.apache.kafka.streams.kstream.TimeWindows
-
Returns a window definition with the given window size, and with the advance interval being
equal to the window size.
- of(String) - Static method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return an unlimited window starting at timestamp zero.
- offset() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The position of this record in the corresponding Kafka partition.
- offset() - Method in class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
- offset() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The offset of the record in the topic/partition.
- offset(Map<TopicPartition, Long>) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Reset the consumer offsets for the given topic partitions.
- offset(TopicPartition, long) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Reset the consumer offsets for the given topic partition.
- offset(Map<String, T>) - Method in interface org.apache.kafka.connect.storage.OffsetStorageReader
-
Get the offset for the specified partition.
- offset() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the offset of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- OffsetAndMetadata - Class in org.apache.kafka.clients.consumer
-
The Kafka offset commit API allows users to provide additional metadata (in the form of a string)
when an offset is committed.
- OffsetAndMetadata(long, String) - Constructor for class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
Construct a new OffsetAndMetadata object for committing through
KafkaConsumer
.
- OffsetAndMetadata(long) - Constructor for class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
Construct a new OffsetAndMetadata object for committing through
KafkaConsumer
.
- OffsetCommitCallback - Interface in org.apache.kafka.clients.consumer
-
A callback interface that the user can implement to trigger custom actions when a commit request completes.
- OffsetMetadataTooLarge - Exception in org.apache.kafka.common.errors
-
The client has tried to save its offset with associated metadata larger than the maximum size allowed by the server.
- OffsetMetadataTooLarge() - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(String) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetOutOfRangeException - Exception in org.apache.kafka.clients.consumer
-
No reset policy has been defined, and the offsets for these partitions are either larger or smaller
than the range of offsets the server has for the given partition.
- OffsetOutOfRangeException(Map<TopicPartition, Long>) - Constructor for exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- OffsetOutOfRangeException - Exception in org.apache.kafka.common.errors
-
No reset policy has been defined, and the offsets for these partitions are either larger or smaller
than the range of offsets the server has for the given partition.
- OffsetOutOfRangeException(String) - Constructor for exception org.apache.kafka.common.errors.OffsetOutOfRangeException
-
- OffsetOutOfRangeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetOutOfRangeException
-
- offsetOutOfRangePartitions() - Method in exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- OffsetResetStrategy - Enum in org.apache.kafka.clients.consumer
-
- offsets(Collection<Map<String, T>>) - Method in interface org.apache.kafka.connect.storage.OffsetStorageReader
-
Get a set of offsets for the specified partition identifiers.
- offsetStorageReader() - Method in interface org.apache.kafka.connect.source.SourceTaskContext
-
Get the OffsetStorageReader for this SourceTask.
- OffsetStorageReader - Interface in org.apache.kafka.connect.storage
-
OffsetStorageReader provides access to the offset storage used by sources.
- onAcknowledgement(RecordMetadata, Exception) - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
This method is called when the record sent to the server has been acknowledged, or when sending the record fails before
it gets sent to the server.
- onCommit(Map<TopicPartition, OffsetAndMetadata>) - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
This is called when offsets get committed.
- onComplete(Map<TopicPartition, OffsetAndMetadata>, Exception) - Method in interface org.apache.kafka.clients.consumer.OffsetCommitCallback
-
A callback method the user can implement to provide asynchronous handling of commit request completion.
- onCompletion(RecordMetadata, Exception) - Method in interface org.apache.kafka.clients.producer.Callback
-
A callback method the user can implement to provide asynchronous handling of request completion.
- onConsume(ConsumerRecords<K, V>) - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
- onPartitionsAssigned(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.ConsumerRebalanceListener
-
A callback method the user can implement to provide handling of customized offsets on completion of a successful
partition re-assignment.
- onPartitionsAssigned(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- onPartitionsRevoked(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.ConsumerRebalanceListener
-
A callback method the user can implement to provide handling of offset commits to a customized store on the start
of a rebalance operation.
- onPartitionsRevoked(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- onSend(ProducerRecord<K, V>) - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
- open(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
The SinkTask use this method to create writers for newly assigned partitions in case of partition
rebalance.
- optional() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set this schema as optional.
- OPTIONAL_BOOLEAN_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_BYTES_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_FLOAT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_FLOAT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT16_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT8_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_STRING_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- orderInGroup - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- org.apache.kafka.clients.consumer - package org.apache.kafka.clients.consumer
-
- org.apache.kafka.clients.producer - package org.apache.kafka.clients.producer
-
- org.apache.kafka.common - package org.apache.kafka.common
-
- org.apache.kafka.common.config - package org.apache.kafka.common.config
-
- org.apache.kafka.common.errors - package org.apache.kafka.common.errors
-
- org.apache.kafka.common.serialization - package org.apache.kafka.common.serialization
-
- org.apache.kafka.connect.connector - package org.apache.kafka.connect.connector
-
- org.apache.kafka.connect.data - package org.apache.kafka.connect.data
-
- org.apache.kafka.connect.errors - package org.apache.kafka.connect.errors
-
- org.apache.kafka.connect.sink - package org.apache.kafka.connect.sink
-
- org.apache.kafka.connect.source - package org.apache.kafka.connect.source
-
- org.apache.kafka.connect.storage - package org.apache.kafka.connect.storage
-
- org.apache.kafka.connect.util - package org.apache.kafka.connect.util
-
- org.apache.kafka.streams - package org.apache.kafka.streams
-
- org.apache.kafka.streams.errors - package org.apache.kafka.streams.errors
-
- org.apache.kafka.streams.kstream - package org.apache.kafka.streams.kstream
-
- org.apache.kafka.streams.processor - package org.apache.kafka.streams.processor
-
- org.apache.kafka.streams.state - package org.apache.kafka.streams.state
-
- originals() - Method in class org.apache.kafka.common.config.AbstractConfig
-
- originalsStrings() - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get all the original settings, ensuring that all values are of type String.
- originalsWithPrefix(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Gets all original settings with the given prefix, stripping the prefix before adding it to the output.
- outerJoin(KStream<K, V1>, ValueJoiner<V, V1, R>, JoinWindows, Serde<K>, Serde<V>, Serde<V1>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream with another
KStream
's elements of the same key using windowed Outer Join.
- outerJoin(KStream<K, V1>, ValueJoiner<V, V1, R>, JoinWindows) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream with another
KStream
's elements of the same key using windowed Outer Join
with default serializers and deserializers.
- outerJoin(KTable<K, V1>, ValueJoiner<V, V1, R>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Combine values of this stream with another
KTable
stream's elements of the same key using Outer Join.
- overlap(Window) - Method in class org.apache.kafka.streams.kstream.Window
-
Check if the given window overlaps with this window.
- pair(K, V) - Static method in class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- parameter(String, String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set a schema parameter.
- parameters() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- parameters() - Method in interface org.apache.kafka.connect.data.Schema
-
Get a map of schema parameters.
- parameters() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- parameters(Map<String, String>) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set schema parameters.
- parse(Map<?, ?>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Parse and validate configs against this configuration definition.
- parse(String) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- partition() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The partition from which this record is received
- partition() - Method in exception org.apache.kafka.clients.consumer.NoOffsetForPartitionException
-
- partition(String, Object, byte[], Object, byte[], Cluster) - Method in interface org.apache.kafka.clients.producer.Partitioner
-
Compute the partition for the given record.
- partition() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- partition() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The partition the record was sent to
- partition(TopicPartition) - Method in class org.apache.kafka.common.Cluster
-
Get the metadata for the specified partition
- partition() - Method in class org.apache.kafka.common.PartitionInfo
-
The partition id
- partition() - Method in class org.apache.kafka.common.TopicPartition
-
- partition() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the partition id of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- partition(K, V, int) - Method in interface org.apache.kafka.streams.processor.StreamPartitioner
-
Determine the partition number for a record with the given key and value and the current number of partitions.
- partition - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the partition.
- PARTITION_ASSIGNMENT_STRATEGY_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
partition.assignment.strategy
- PARTITION_GROUPER_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
partition.grouper
- partitionCountForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the number of partitions for the given topic
- Partitioner - Interface in org.apache.kafka.clients.producer
-
Partitioner Interface
- PARTITIONER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
partitioner.class
- PartitionGrouper - Interface in org.apache.kafka.streams.processor
-
A partition grouper that generates partition groups given the list of topic-partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
Generate tasks with the assigned topic partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in interface org.apache.kafka.streams.processor.PartitionGrouper
-
Returns a map of task ids to groups of partitions.
- PartitionInfo - Class in org.apache.kafka.common
-
Information about a topic-partition.
- PartitionInfo(String, int, Node, Node[], Node[]) - Constructor for class org.apache.kafka.common.PartitionInfo
-
- partitions() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get the partitions which have records contained in this record set.
- partitions() - Method in exception org.apache.kafka.clients.consumer.InvalidOffsetException
-
- partitions() - Method in exception org.apache.kafka.clients.consumer.NoOffsetForPartitionException
-
- partitions() - Method in exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- partitionsFor(String) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- partitionsFor(String) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get metadata about the partitions for a given topic.
- partitionsFor(String) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- partitionsFor(String) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Get the partition metadata for the give topic.
- partitionsFor(String) - Method in class org.apache.kafka.clients.producer.MockProducer
-
- partitionsFor(String) - Method in interface org.apache.kafka.clients.producer.Producer
-
Get a list of partitions for the given topic for custom partition assignment.
- partitionsForNode(int) - Method in class org.apache.kafka.common.Cluster
-
Get the list of partitions whose leader is this node
- partitionsForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the list of partitions for this topic
- pause(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- pause(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Suspend fetching from the requested partitions.
- pause(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- pause(TopicPartition...) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Pause consumption of messages from the specified TopicPartitions.
- paused() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- paused() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
- paused() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- persistent() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Return if the storage is persistent or not.
- persistent() - Method in interface org.apache.kafka.streams.state.Stores.KeyValueFactory
-
Keep all key-value entries off-heap in a local database, although for durability all entries are recorded in a Kafka
topic that can be read to restore the entries if they are lost.
- poll(long) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- poll(long) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Fetch data for the topics or partitions specified using one of the subscribe/assign APIs.
- poll(long) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- poll() - Method in class org.apache.kafka.connect.source.SourceTask
-
Poll this SourceTask for new records.
- POLL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
poll.ms
- port() - Method in class org.apache.kafka.common.Node
-
The port for this node
- position(TopicPartition) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- position(TopicPartition) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the offset of the next record that will be fetched (if a record with that offset exists).
- position(TopicPartition) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- Predicate<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
Predicate
interface represents a predicate (boolean-valued function) of a key-value pair.
- PRINCIPAL_BUILDER_CLASS_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- PRINCIPAL_BUILDER_CLASS_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- print() - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the elements of this stream to System.out
Implementors will need to override toString for keys and values that are not of
type String, Integer etc to get meaningful information.
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the elements of this stream to System.out
- print() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the elements of this stream to System.out
Implementors will need to override toString for keys and values that are not of
type String, Integer etc to get meaningful information.
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the elements of this stream to System.out
- process(ProcessorSupplier<K, V>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Process all elements in this stream, one element at a time, by applying a
Processor
.
- process(K, V) - Method in interface org.apache.kafka.streams.processor.Processor
-
Process the record with the given key and value.
- Processor<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor of key-value pair records.
- ProcessorContext - Interface in org.apache.kafka.streams.processor
-
Processor context interface.
- ProcessorStateException - Exception in org.apache.kafka.streams.errors
-
Indicates a processor state operation (e.g.
- ProcessorStateException(String) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorSupplier<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor supplier that can create one or more
Processor
instances.
- Producer<K,V> - Interface in org.apache.kafka.clients.producer
-
- ProducerConfig - Class in org.apache.kafka.clients.producer
-
Configuration for the Kafka Producer.
- ProducerInterceptor<K,V> - Interface in org.apache.kafka.clients.producer
-
A plugin interface that allows you to intercept (and possibly mutate) the records received by the producer before
they are published to the Kafka cluster.
- ProducerRecord<K,V> - Class in org.apache.kafka.clients.producer
-
A key/value pair to be sent to Kafka.
- ProducerRecord(String, Integer, Long, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Creates a record with a specified timestamp to be sent to a specified topic and partition
- ProducerRecord(String, Integer, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Creates a record to be sent to a specified topic and partition
- ProducerRecord(String, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Create a record to be sent to Kafka
- ProducerRecord(String, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Create a record with no key
- project(Schema, Object, Schema) - Static method in class org.apache.kafka.connect.data.SchemaProjector
-
This method project a value between compatible schemas and throw exceptions when non compatible schemas are provided
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
- punctuate(long) - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- punctuate(long) - Method in interface org.apache.kafka.streams.processor.Processor
-
- put(String, Object) - Method in class org.apache.kafka.connect.data.Struct
-
Set the value of a field.
- put(Field, Object) - Method in class org.apache.kafka.connect.data.Struct
-
Set the value of a field.
- put(Collection<SinkRecord>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Put the records in the sink.
- put(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key
- put(K, V) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the current wall-clock time as the timestamp
into the corresponding window
- put(K, V, long) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the given timestamp into the corresponding window
- putAll(List<KeyValue<K, V>>) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update all the given key/value pairs
- putIfAbsent(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key, unless a value
is already associated with the key
- rack() - Method in class org.apache.kafka.common.Node
-
The rack for this node
- raiseError(Exception) - Method in interface org.apache.kafka.connect.connector.ConnectorContext
-
Raise an unrecoverable exception to the Connect framework.
- range(K, K) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Get an iterator over a given range of keys.
- RangeAssignor - Class in org.apache.kafka.clients.consumer
-
The range assignor works on a per-topic basis.
- RangeAssignor() - Constructor for class org.apache.kafka.clients.consumer.RangeAssignor
-
- rawKey(K) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given key.
- rawValue(V) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given value.
- readFrom(DataInputStream) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- readFrom(ByteBuffer) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- rebalance(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
Simulate a rebalance event.
- RebalanceInProgressException - Exception in org.apache.kafka.common.errors
-
- RebalanceInProgressException() - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(String) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
receive.buffer.bytes
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
receive.buffer.bytes
- recommendedValues() - Method in class org.apache.kafka.common.config.ConfigValue
-
- recommendedValues(List<Object>) - Method in class org.apache.kafka.common.config.ConfigValue
-
- recommender - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- reconfigure(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Reconfigure this Connector.
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
reconnect.backoff.ms
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
reconnect.backoff.ms
- RecordBatchTooLargeException - Exception in org.apache.kafka.common.errors
-
This record batch is larger than the maximum allowable size
- RecordBatchTooLargeException() - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(String) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- recordLatency(Sensor, long, long) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
- RecordMetadata - Class in org.apache.kafka.clients.producer
-
The metadata for a record that has been acknowledged by the server
- RecordMetadata(TopicPartition, long, long) - Constructor for class org.apache.kafka.clients.producer.RecordMetadata
-
Deprecated.
- RecordMetadata(TopicPartition, long, long, long, long, int, int) - Constructor for class org.apache.kafka.clients.producer.RecordMetadata
-
- records(TopicPartition) - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get just the records for the given partition
- records(String) - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get just the records for the given topic
- RecordTooLargeException - Exception in org.apache.kafka.common.errors
-
This record is larger than the maximum allowable size
- RecordTooLargeException() - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String, Map<TopicPartition, Long>) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- recordTooLargePartitions() - Method in exception org.apache.kafka.common.errors.RecordTooLargeException
-
- reduce(Reducer<V>, Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine updating values of this stream by the selected key into a new instance of
KTable
.
- reduceByKey(Reducer<V>, Windows<W>, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream by key on a window basis into a new instance of windowed
KTable
.
- reduceByKey(Reducer<V>, Windows<W>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream by key on a window basis into a new instance of windowed
KTable
with default serializers and deserializers.
- reduceByKey(Reducer<V>, Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream by key into a new instance of ever-updating
KTable
.
- reduceByKey(Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Combine values of this stream by key into a new instance of ever-updating
KTable
with default serializers and deserializers.
- Reducer<V> - Interface in org.apache.kafka.streams.kstream
-
The
Reducer
interface for combining two values of the same type into a new value.
- register(StateStore, boolean, StateRestoreCallback) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Registers and possibly restores the specified storage engine.
- ReplicaNotAvailableException - Exception in org.apache.kafka.common.errors
-
- ReplicaNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- ReplicaNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- ReplicaNotAvailableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- replicas() - Method in class org.apache.kafka.common.PartitionInfo
-
The complete set of replicas for this partition regardless of whether they are alive or up-to-date
- REPLICATION_FACTOR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
replication.factor
- REPLICATION_FACTOR_DOC - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
request.timeout.ms
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
request.timeout.ms
- requestTaskReconfiguration() - Method in interface org.apache.kafka.connect.connector.ConnectorContext
-
Requests that the runtime reconfigure the Tasks for this source.
- required() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set this schema as required.
- restore(byte[], byte[]) - Method in interface org.apache.kafka.streams.processor.StateRestoreCallback
-
- resume(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- resume(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
- resume(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- resume(TopicPartition...) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Resume consumption of messages from previously paused TopicPartitions.
- RetriableCommitFailedException - Exception in org.apache.kafka.clients.consumer
-
- RetriableCommitFailedException(String) - Constructor for exception org.apache.kafka.clients.consumer.RetriableCommitFailedException
-
- RetriableCommitFailedException(String, Throwable) - Constructor for exception org.apache.kafka.clients.consumer.RetriableCommitFailedException
-
- RetriableException - Exception in org.apache.kafka.common.errors
-
A retryable exception is a transient exception that if retried may succeed.
- RetriableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException(String) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException() - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException - Exception in org.apache.kafka.connect.errors
-
An exception that indicates the operation can be reattempted.
- RetriableException(String) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RetriableException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RetriableException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RETRIES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
retries
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
retry.backoff.ms
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
retry.backoff.ms
- RoundRobinAssignor - Class in org.apache.kafka.clients.consumer
-
The roundrobin assignor lays out all the available partitions and all the available consumers.
- RoundRobinAssignor() - Constructor for class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- SASL_ENABLED_MECHANISMS - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_ENABLED_MECHANISMS_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_KINIT_CMD - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_KINIT_CMD_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_MIN_TIME_BEFORE_RELOGIN - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_MIN_TIME_BEFORE_RELOGIN_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_SERVICE_NAME - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_SERVICE_NAME_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_JITTER - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_JITTER_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_WINDOW_FACTOR - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_WINDOW_FACTOR_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
SASL mechanism configuration - standard mechanism names are listed
here.
- SASL_MECHANISM_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SaslConfigs - Class in org.apache.kafka.common.config
-
- SaslConfigs() - Constructor for class org.apache.kafka.common.config.SaslConfigs
-
- SCALE_FIELD - Static variable in class org.apache.kafka.connect.data.Decimal
-
- schedule(long) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Schedules a periodic operation for processors.
- scheduleNopPollTask() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- schedulePollTask(Runnable) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
Schedule a task to be executed during a poll().
- schema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Date
-
- schema(int) - Static method in class org.apache.kafka.connect.data.Decimal
-
- schema() - Method in class org.apache.kafka.connect.data.Field
-
Get the schema of this field
- Schema - Interface in org.apache.kafka.connect.data
-
Definition of an abstract data type.
- schema() - Method in interface org.apache.kafka.connect.data.Schema
-
Return a concrete instance of the
Schema
- schema() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- schema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Return a concrete instance of the
Schema
specified by this builder
- schema() - Method in class org.apache.kafka.connect.data.Struct
-
Get the schema for this Struct.
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Time
-
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Timestamp
-
- Schema.Type - Enum in org.apache.kafka.connect.data
-
The type of a schema.
- SchemaAndValue - Class in org.apache.kafka.connect.data
-
- SchemaAndValue(Schema, Object) - Constructor for class org.apache.kafka.connect.data.SchemaAndValue
-
- SchemaBuilder - Class in org.apache.kafka.connect.data
-
SchemaBuilder provides a fluent API for constructing
Schema
objects.
- SchemaBuilderException - Exception in org.apache.kafka.connect.errors
-
- SchemaBuilderException(String) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaBuilderException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaBuilderException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaProjector - Class in org.apache.kafka.connect.data
-
SchemaProjector is utility to project a value between compatible schemas and throw exceptions
when non compatible schemas are provided.
- SchemaProjector() - Constructor for class org.apache.kafka.connect.data.SchemaProjector
-
- SchemaProjectorException - Exception in org.apache.kafka.connect.errors
-
- SchemaProjectorException(String) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- SchemaProjectorException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- SchemaProjectorException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- schemaType(Class<?>) - Static method in class org.apache.kafka.connect.data.ConnectSchema
-
- seek(TopicPartition, long) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seek(TopicPartition, long) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Overrides the fetch offsets that the consumer will use on the next
poll(timeout)
.
- seek(TopicPartition, long) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- seekToBeginning(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seekToBeginning(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Seek to the first offset for each of the given partitions.
- seekToBeginning(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- seekToEnd(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seekToEnd(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Seek to the last offset for each of the given partitions.
- seekToEnd(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- segments - Variable in class org.apache.kafka.streams.kstream.Windows
-
- segments(int) - Method in class org.apache.kafka.streams.kstream.Windows
-
Specify the number of segments to be used for rolling the window store,
this function is not exposed to users but can be called by developers that extend this JoinWindows specs.
- selectKey(KeyValueMapper<K, V, K1>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new key from the current key and value.
- send(ProducerRecord<K, V>) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Asynchronously send a record to a topic.
- send(ProducerRecord<K, V>, Callback) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Asynchronously send a record to a topic and invoke the provided callback when the send has been acknowledged.
- send(ProducerRecord<K, V>) - Method in class org.apache.kafka.clients.producer.MockProducer
-
Adds the record to the list of sent records.
- send(ProducerRecord<K, V>, Callback) - Method in class org.apache.kafka.clients.producer.MockProducer
-
Adds the record to the list of sent records.
- send(ProducerRecord<K, V>) - Method in interface org.apache.kafka.clients.producer.Producer
-
Send the given record asynchronously and return a future which will eventually contain the response information.
- send(ProducerRecord<K, V>, Callback) - Method in interface org.apache.kafka.clients.producer.Producer
-
Send a record and invoke the given callback when the record has been acknowledged by the server
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
send.buffer.bytes
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
send.buffer.bytes
- Serde<T> - Interface in org.apache.kafka.common.serialization
-
The interface for wrapping a serializer and deserializer for the given data type.
- serdeFrom(Class<T>) - Static method in class org.apache.kafka.common.serialization.Serdes
-
- serdeFrom(Serializer<T>, Deserializer<T>) - Static method in class org.apache.kafka.common.serialization.Serdes
-
Construct a serde object from separate serializer and deserializer
- Serdes - Class in org.apache.kafka.common.serialization
-
Factory for creating serializers / deserializers.
- Serdes() - Constructor for class org.apache.kafka.common.serialization.Serdes
-
- Serdes.ByteArraySerde - Class in org.apache.kafka.common.serialization
-
- Serdes.ByteArraySerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.ByteArraySerde
-
- Serdes.ByteBufferSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.ByteBufferSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.ByteBufferSerde
-
- Serdes.BytesSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.BytesSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.BytesSerde
-
- Serdes.DoubleSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.DoubleSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.DoubleSerde
-
- Serdes.IntegerSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.IntegerSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.IntegerSerde
-
- Serdes.LongSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.LongSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.LongSerde
-
- Serdes.StringSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.StringSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.StringSerde
-
- SerializationException - Exception in org.apache.kafka.common.errors
-
Any exception during serialization in the producer
- SerializationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException(String) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException(Throwable) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException() - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- serialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- serialize(String, ByteBuffer) - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- serialize(String, Bytes) - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- serialize(String, Double) - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- serialize(String, Integer) - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- serialize(String, Long) - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- serialize(String, T) - Method in interface org.apache.kafka.common.serialization.Serializer
-
- serialize(String, String) - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- serializedKeySize() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The size of the serialized, uncompressed key in bytes.
- serializedKeySize() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The size of the serialized, uncompressed key in bytes.
- serializedValueSize() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The size of the serialized, uncompressed value in bytes.
- serializedValueSize() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The size of the serialized, uncompressed value in bytes.
- serializer() - Method in interface org.apache.kafka.common.serialization.Serde
-
- Serializer<T> - Interface in org.apache.kafka.common.serialization
-
- SESSION_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
session.timeout.ms
- setException(KafkaException) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- setUncaughtExceptionHandler(Thread.UncaughtExceptionHandler) - Method in class org.apache.kafka.streams.KafkaStreams
-
Sets the handler invoked when a stream thread abruptly terminates due to an uncaught exception.
- SinkConnector - Class in org.apache.kafka.connect.sink
-
SinkConnectors implement the Connector interface to send Kafka data to another system.
- SinkConnector() - Constructor for class org.apache.kafka.connect.sink.SinkConnector
-
- SinkRecord - Class in org.apache.kafka.connect.sink
-
SinkRecord is a
ConnectRecord
that has been read from Kafka and includes the kafkaOffset of
the record in the Kafka topic-partition in addition to the standard fields.
- SinkRecord(String, int, Schema, Object, Schema, Object, long) - Constructor for class org.apache.kafka.connect.sink.SinkRecord
-
- SinkTask - Class in org.apache.kafka.connect.sink
-
SinkTask is a Task that takes records loaded from Kafka and sends them to another system.
- SinkTask() - Constructor for class org.apache.kafka.connect.sink.SinkTask
-
- SinkTaskContext - Interface in org.apache.kafka.connect.sink
-
Context passed to SinkTasks, allowing them to access utilities in the Kafka Connect runtime.
- sinkTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- size - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the window, i.e.
- SourceConnector - Class in org.apache.kafka.connect.source
-
SourceConnectors implement the connector interface to pull data from another system and send
it to Kafka.
- SourceConnector() - Constructor for class org.apache.kafka.connect.source.SourceConnector
-
- sourceOffset() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- sourcePartition() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord - Class in org.apache.kafka.connect.source
-
SourceRecords are generated by SourceTasks and passed to Kafka Connect for storage in
Kafka.
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Integer, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Schema, Object, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Integer, Schema, Object, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceTask - Class in org.apache.kafka.connect.source
-
SourceTask is a Task that pulls records from another system for storage in Kafka.
- SourceTask() - Constructor for class org.apache.kafka.connect.source.SourceTask
-
- SourceTaskContext - Interface in org.apache.kafka.connect.source
-
SourceTaskContext is provided to SourceTasks to allow them to interact with the underlying
runtime.
- sourceTopics(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Get the names of topics that are to be consumed by the source nodes created by this builder.
- sourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- SSL_CIPHER_SUITES_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CIPHER_SUITES_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CLIENT_AUTH_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CLIENT_AUTH_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENABLED_PROTOCOLS_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENABLED_PROTOCOLS_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENDPOINT_IDENTIFICATION_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENDPOINT_IDENTIFICATION_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEY_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEY_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYMANAGER_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYMANAGER_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_LOCATION_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_LOCATION_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_TYPE_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_TYPE_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROTOCOL_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROTOCOL_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROVIDER_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROVIDER_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTMANAGER_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTMANAGER_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_LOCATION_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_LOCATION_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_TYPE_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_TYPE_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SslConfigs - Class in org.apache.kafka.common.config
-
- SslConfigs() - Constructor for class org.apache.kafka.common.config.SslConfigs
-
- start(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Start this Connector.
- start(Map<String, String>) - Method in interface org.apache.kafka.connect.connector.Task
-
Start the Task
- start(Map<String, String>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Start the Task.
- start(Map<String, String>) - Method in class org.apache.kafka.connect.source.SourceTask
-
Start the Task.
- start() - Method in class org.apache.kafka.streams.KafkaStreams
-
Start the stream instance by starting all its threads.
- start - Variable in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
The start timestamp of the window.
- start() - Method in class org.apache.kafka.streams.kstream.Window
-
Return the start timestamp of this window, inclusive
- startOn(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return a new unlimited window for the specified start timestamp.
- STATE_CLEANUP_DELAY_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.cleanup.delay
- STATE_DIR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.dir
- stateChangelogTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- stateDir() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the state directory for the partition.
- stateName() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the name of the state.
- StateRestoreCallback - Interface in org.apache.kafka.streams.processor
-
Restoration logic for log-backed state stores upon restart,
it takes one record at a time from the logs to apply to the restoring state.
- StateSerdes<K,V> - Class in org.apache.kafka.streams.state
-
Factory for creating serializers / deserializers for state stores in Kafka Streams.
- StateSerdes(String, Serde<K>, Serde<V>) - Constructor for class org.apache.kafka.streams.state.StateSerdes
-
Create a context for serialization using the specified serializers and deserializers which
must match the key and value types used as parameters for this object; the state changelog topic
is provided to bind this serde factory to, so that future calls for serialize / deserialize do not
need to provide the topic name any more.
- StateStore - Interface in org.apache.kafka.streams.processor
-
A storage engine for managing state maintained by a stream processor.
- StateStoreSupplier - Interface in org.apache.kafka.streams.processor
-
A state store supplier which can create one or more
StateStore
instances.
- stop() - Method in class org.apache.kafka.connect.connector.Connector
-
Stop this connector.
- stop() - Method in interface org.apache.kafka.connect.connector.Task
-
Stop this task.
- stop() - Method in class org.apache.kafka.connect.sink.SinkTask
-
Perform any cleanup to stop this task.
- stop() - Method in class org.apache.kafka.connect.source.SourceTask
-
Signal this SourceTask to stop.
- Stores - Class in org.apache.kafka.streams.state
-
Factory for creating state stores in Kafka Streams.
- Stores() - Constructor for class org.apache.kafka.streams.state.Stores
-
- Stores.InMemoryKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create in-memory key-value stores.
- Stores.KeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to specify the different kinds of key-value stores.
- Stores.PersistentKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create off-heap key-value stores that use a local database.
- Stores.StoreFactory - Class in org.apache.kafka.streams.state
-
- Stores.StoreFactory() - Constructor for class org.apache.kafka.streams.state.Stores.StoreFactory
-
- Stores.ValueFactory<K> - Class in org.apache.kafka.streams.state
-
The factory for creating off-heap key-value stores.
- Stores.ValueFactory() - Constructor for class org.apache.kafka.streams.state.Stores.ValueFactory
-
- stream(String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
instance from the specified topics.
- stream(Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
instance from the specified topics.
- STREAM_THREAD_INSTANCE - Static variable in class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamPartitioner<K,V> - Interface in org.apache.kafka.streams.processor
-
Determine how records are distributed among the partitions in a Kafka topic.
- StreamsConfig - Class in org.apache.kafka.streams
-
Configuration for Kafka Streams.
- StreamsConfig(Map<?, ?>) - Constructor for class org.apache.kafka.streams.StreamsConfig
-
- StreamsConfig.InternalConfig - Class in org.apache.kafka.streams
-
- StreamsConfig.InternalConfig() - Constructor for class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamsException - Exception in org.apache.kafka.streams.errors
-
StreamsException is the top-level exception type generated by Kafka Streams.
- StreamsException(String) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsMetrics - Interface in org.apache.kafka.streams
-
The Kafka Streams metrics interface for adding metric sensors and collecting metric values.
- String() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- string() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- STRING_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- StringConverter - Class in org.apache.kafka.connect.storage
-
Converter
implementation that only supports serializing to strings.
- StringConverter() - Constructor for class org.apache.kafka.connect.storage.StringConverter
-
- StringDeserializer - Class in org.apache.kafka.common.serialization
-
String encoding defaults to UTF8 and can be customized by setting the property key.deserializer.encoding,
value.deserializer.encoding or deserializer.encoding.
- StringDeserializer() - Constructor for class org.apache.kafka.common.serialization.StringDeserializer
-
- StringSerializer - Class in org.apache.kafka.common.serialization
-
String encoding defaults to UTF8 and can be customized by setting the property key.serializer.encoding,
value.serializer.encoding or serializer.encoding.
- StringSerializer() - Constructor for class org.apache.kafka.common.serialization.StringSerializer
-
- struct() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- Struct - Class in org.apache.kafka.connect.data
-
A structured record containing a set of named fields with values, each field using an independent
Schema
.
- Struct(Schema) - Constructor for class org.apache.kafka.connect.data.Struct
-
Create a new Struct for this
Schema
- subscribe(Collection<String>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Pattern, ConsumerRebalanceListener) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to the given list of topics to get dynamically
assigned partitions.
- subscribe(Collection<String>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to the given list of topics to get dynamically assigned partitions.
- subscribe(Pattern, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to all topics matching specified pattern to get dynamically assigned partitions.
- subscribe(Collection<String>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscribe(Pattern, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscription() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscription() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the current subscription.
- subscription() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- table(String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
instance for the specified topic.
- table(Serde<K>, Serde<V>, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
instance for the specified topic.
- tags() - Method in class org.apache.kafka.common.MetricName
-
- Task - Interface in org.apache.kafka.connect.connector
-
Tasks contain the code that actually copies data to/from another system.
- TaskAssignmentException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying to assign stream tasks to threads
- TaskAssignmentException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- taskClass() - Method in class org.apache.kafka.connect.connector.Connector
-
Returns the Task implementation for this Connector.
- taskConfigs(int) - Method in class org.apache.kafka.connect.connector.Connector
-
Returns a set of configurations for Tasks based on the current configuration,
producing at most count configurations.
- taskId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the task id
- TaskId - Class in org.apache.kafka.streams.processor
-
The task ID representation composed as topic group ID plus the assigned partition ID.
- TaskId(int, int) - Constructor for class org.apache.kafka.streams.processor.TaskId
-
- TaskIdFormatException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying parse the task id from the read string
- TaskIdFormatException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- test(K, V) - Method in interface org.apache.kafka.streams.kstream.Predicate
-
Test if the record with the given key and value satisfies the predicate.
- through(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, also creates a new instance of
KStream
from the topic
using default serializers and deserializers and producer's
DefaultPartitioner
.
- through(StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, also creates a new instance of
KStream
from the topic
using default serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, also creates a new instance of
KStream
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, also creates a new instance of
KStream
from the topic
using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic, also creates a new instance of
KTable
from the topic
using default serializers and deserializers and producer's
DefaultPartitioner
.
- through(StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic, also creates a new instance of
KTable
from the topic using default serializers
and deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic, also creates a new instance of
KTable
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic, also creates a new instance of
KTable
from the topic
using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- Time - Class in org.apache.kafka.connect.data
-
A time representing a specific point in a day, not tied to any specific date.
- Time() - Constructor for class org.apache.kafka.connect.data.Time
-
- timeout(long) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Set the timeout in milliseconds.
- TIMEOUT_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- TimeoutException - Exception in org.apache.kafka.common.errors
-
Indicates that a request timed out.
- TimeoutException() - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(String) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(Throwable) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- timestamp() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The timestamp of this record
- timestamp() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- timestamp() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The timestamp of the record in the topic/partition.
- Timestamp - Class in org.apache.kafka.connect.data
-
A timestamp representing an absolute time, without timezone information.
- Timestamp() - Constructor for class org.apache.kafka.connect.data.Timestamp
-
- timestamp() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the current timestamp.
- TIMESTAMP_EXTRACTOR_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
timestamp.extractor
- TimestampExtractor - Interface in org.apache.kafka.streams.processor
-
An interface that allows the Kafka Streams framework to extract a timestamp from an instance of
ConsumerRecord
.
- timestampType() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The timestamp type of this record
- TimeWindows - Class in org.apache.kafka.streams.kstream
-
The time-based window specifications used for aggregations.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config
and producer's DefaultPartitioner
.
- to(StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic using default serializers specified in the config
and producer's DefaultPartitioner
.
- to(StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic using default serializers specified in the config
and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<K, V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this stream to a topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- toConnectData(String, byte[]) - Method in interface org.apache.kafka.connect.storage.Converter
-
Convert a native object to a Kafka Connect data object.
- toConnectData(String, byte[]) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- toHtmlTable() - Method in class org.apache.kafka.common.config.ConfigDef
-
- toLogical(Schema, int) - Static method in class org.apache.kafka.connect.data.Date
-
- toLogical(Schema, byte[]) - Static method in class org.apache.kafka.connect.data.Decimal
-
- toLogical(Schema, int) - Static method in class org.apache.kafka.connect.data.Time
-
- toLogical(Schema, long) - Static method in class org.apache.kafka.connect.data.Timestamp
-
- topic() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The topic this record is received from
- topic() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- topic() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The topic the record was appended to
- topic() - Method in class org.apache.kafka.common.PartitionInfo
-
The topic name
- topic() - Method in class org.apache.kafka.common.TopicPartition
-
- topic() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- topic() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the topic name of the current input record; could be null if it is not
available (for example, if this method is invoked from the punctuate call)
- TopicAuthorizationException - Exception in org.apache.kafka.common.errors
-
- TopicAuthorizationException(Set<String>) - Constructor for exception org.apache.kafka.common.errors.TopicAuthorizationException
-
- TopicAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.TopicAuthorizationException
-
- topicGroupId - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the topic group.
- topicGroups(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the map of topic groups keyed by the group id.
- TopicPartition - Class in org.apache.kafka.common
-
A topic name and partition number
- TopicPartition(String, int) - Constructor for class org.apache.kafka.common.TopicPartition
-
- topics() - Method in class org.apache.kafka.common.Cluster
-
Get all topics.
- TOPICS_CONFIG - Static variable in class org.apache.kafka.connect.sink.SinkConnector
-
Configuration key for the list of input topics for this connector.
- TOPICS_CONFIG - Static variable in class org.apache.kafka.connect.sink.SinkTask
-
The configuration key that provides the list of topics that are inputs for this
SinkTask.
- TopologyBuilder - Class in org.apache.kafka.streams.processor
-
A component that is used to build a ProcessorTopology
.
- TopologyBuilder() - Constructor for class org.apache.kafka.streams.processor.TopologyBuilder
-
Create a new builder.
- TopologyBuilder.TopicsInfo - Class in org.apache.kafka.streams.processor
-
- TopologyBuilder.TopicsInfo(Set<String>, Set<String>, Set<String>, Set<String>) - Constructor for class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- TopologyBuilderException - Exception in org.apache.kafka.streams.errors
-
Indicates a pre-run time error incurred while parsing the builder to construct the processor topology
- TopologyBuilderException(String) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- toRst() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the configs formatted with reStructuredText, suitable for embedding in Sphinx
documentation.
- toStream() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Convert this stream to a new instance of
KStream
.
- toStream(KeyValueMapper<K, V, K1>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- toString() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
- toString() - Method in class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
- toString() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- toString() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
- toString() - Method in class org.apache.kafka.common.Cluster
-
- toString() - Method in class org.apache.kafka.common.config.ConfigDef.Range
-
- toString() - Method in class org.apache.kafka.common.config.ConfigDef.ValidString
-
- toString() - Method in class org.apache.kafka.common.config.ConfigValue
-
- toString() - Method in class org.apache.kafka.common.MetricName
-
- toString() - Method in class org.apache.kafka.common.Node
-
- toString() - Method in class org.apache.kafka.common.PartitionInfo
-
- toString() - Method in class org.apache.kafka.common.TopicPartition
-
- toString() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- toString() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- toString() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- toString() - Method in class org.apache.kafka.connect.sink.SinkRecord
-
- toString() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- toString() - Method in class org.apache.kafka.streams.KeyValue
-
- toString() - Method in class org.apache.kafka.streams.kstream.Windowed
-
- toString() - Method in class org.apache.kafka.streams.processor.TaskId
-
- transform(TransformerSupplier<K, V, KeyValue<K1, V1>>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new
KStream
instance by applying a
Transformer
to all elements in this stream, one element at a time.
- transform(K, V) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Transform the record with the given key and value.
- transform(V) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Transform the record with the given key and value.
- Transformer<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
A stateful
Transformer
interface for transform a key-value pair into a new value.
- TransformerSupplier<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
- transformValues(ValueTransformerSupplier<V, R>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
- type - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- type() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- type() - Method in interface org.apache.kafka.connect.data.Schema
-
- type() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- type(Schema.Type) - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
Create a SchemaBuilder for the specified type.
- validate(Map<String, String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Validate the current configuration values with the configuration definition.
- validate(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Validate the connector configuration values against configuration definitions.
- validate() - Method in class org.apache.kafka.connect.data.Struct
-
Validates that this struct has filled in all the necessary data with valid values.
- validateValue(Schema, Object) - Static method in class org.apache.kafka.connect.data.ConnectSchema
-
Validate that the value can be used with the schema, i.e.
- validateValue(Object) - Method in class org.apache.kafka.connect.data.ConnectSchema
-
Validate that the value can be used for this schema, i.e.
- validator - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- validValues(String, Map<String, Object>) - Method in interface org.apache.kafka.common.config.ConfigDef.Recommender
-
The valid values for the configuration given the current configuration values.
- value() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The value
- value() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- value() - Method in class org.apache.kafka.common.config.ConfigValue
-
- value(Object) - Method in class org.apache.kafka.common.config.ConfigValue
-
- value() - Method in interface org.apache.kafka.common.Metric
-
The value of the metric
- value() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- value() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- value - Variable in class org.apache.kafka.streams.KeyValue
-
The value of the key-value pair.
- VALUE_DESERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
value.deserializer
- VALUE_DESERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- VALUE_SERDE_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
value.serde
- VALUE_SERDE_CLASS_DOC - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- VALUE_SERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
value.serializer
- VALUE_SERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- valueDeserializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value deserializer.
- valueFrom(byte[]) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Deserialize the value from raw bytes.
- ValueJoiner<V1,V2,R> - Interface in org.apache.kafka.streams.kstream
-
The
ValueJoiner
interface for joining two values into a new value.
- ValueMapper<V1,V2> - Interface in org.apache.kafka.streams.kstream
-
The
ValueMapper
interface for mapping an original value to a new value (could be another key-value pair).
- valueOf(String) - Static method in enum org.apache.kafka.clients.consumer.OffsetResetStrategy
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Importance
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Type
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Width
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.connect.data.Schema.Type
-
Returns the enum constant of this type with the specified name.
- values() - Static method in enum org.apache.kafka.clients.consumer.OffsetResetStrategy
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Method in class org.apache.kafka.common.config.AbstractConfig
-
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Importance
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Type
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Width
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.connect.data.Schema.Type
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- valueSchema() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- valueSchema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- valueSchema() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the value schema for this map or array schema.
- valueSchema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- valueSerde() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the default value serde
- valueSerde() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value serde.
- valueSerde() - Method in class org.apache.kafka.streams.StreamsConfig
-
- valueSerializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value serializer.
- ValueTransformer<V,R> - Interface in org.apache.kafka.streams.kstream
-
- ValueTransformerSupplier<V,R> - Interface in org.apache.kafka.streams.kstream
-
- version() - Method in class org.apache.kafka.connect.connector.Connector
-
Get the version of this connector.
- version() - Method in interface org.apache.kafka.connect.connector.Task
-
Get the version of this task.
- version() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- version() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the optional version of the schema.
- version() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- version(Integer) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the version of this schema.
- visible(String, Map<String, Object>) - Method in interface org.apache.kafka.common.config.ConfigDef.Recommender
-
Set the visibility of the configuration given the current configuration values.
- visible() - Method in class org.apache.kafka.common.config.ConfigValue
-
- visible(boolean) - Method in class org.apache.kafka.common.config.ConfigValue
-
- wakeup() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- wakeup() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Wakeup the consumer.
- wakeup() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- WakeupException - Exception in org.apache.kafka.common.errors
-
Exception used to indicate preemption of a blocking operation by an external thread.
- WakeupException() - Constructor for exception org.apache.kafka.common.errors.WakeupException
-
- WallclockTimestampExtractor - Class in org.apache.kafka.streams.processor
-
- WallclockTimestampExtractor() - Constructor for class org.apache.kafka.streams.processor.WallclockTimestampExtractor
-
- width - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- Window - Class in org.apache.kafka.streams.kstream
-
A single window instance, defined by its start and end timestamp.
- Window(long, long) - Constructor for class org.apache.kafka.streams.kstream.Window
-
Create a new window for the given start time (inclusive) and end time (exclusive).
- window() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the window containing the values associated with this key.
- Windowed<K> - Class in org.apache.kafka.streams.kstream
-
The windowed key interface used in
KTable
, used for representing a windowed table result from windowed stream aggregations,
i.e.
- Windowed(K, Window) - Constructor for class org.apache.kafka.streams.kstream.Windowed
-
- windowed(long, int, boolean) - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Set the persistent store as a windowed key-value store
- Windows<W extends Window> - Class in org.apache.kafka.streams.kstream
-
The window specification interface that can be extended for windowing operation in joins and aggregations.
- Windows(String) - Constructor for class org.apache.kafka.streams.kstream.Windows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.Windows
-
Creates all windows that contain the provided timestamp, indexed by non-negative window start timestamps.
- WindowStore<K,V> - Interface in org.apache.kafka.streams.state
-
- WindowStoreIterator<E> - Interface in org.apache.kafka.streams.state
-
- with(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are within timeDifference
.
- withBuiltinTypes(String, Class<K>, Class<V>) - Static method in class org.apache.kafka.streams.state.StateSerdes
-
Create a new instance of
StateSerdes
for the given state name and key-/value-type classes.
- withByteArrayKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be byte arrays.
- withByteArrayValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use byte arrays for values.
- withByteBufferKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withByteBufferValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withClientSaslSupport() - Method in class org.apache.kafka.common.config.ConfigDef
-
Add standard SASL client configuration options.
- withClientSslSupport() - Method in class org.apache.kafka.common.config.ConfigDef
-
Add standard SSL client configuration options.
- withDoubleKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withDoubleValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withIntegerKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withIntegerValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withKeys(Class<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withKeys(Serde<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the serializer and deserializer for the keys.
- withLongKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withLongValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withPartitions(Map<TopicPartition, PartitionInfo>) - Method in class org.apache.kafka.common.Cluster
-
Return a copy of this cluster combined with `partitions`.
- withStringKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withStringValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withValues(Class<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use values of the specified type.
- withValues(Serde<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use the specified serializer and deserializer for the values.
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the elements of this stream to a file at the given path.
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the elements of this stream to a file at the given path using default serializers and deserializers.
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the elements of this stream to a file at the given path.
- writeTo(DataOutputStream) - Method in class org.apache.kafka.streams.processor.TaskId
-
- writeTo(ByteBuffer) - Method in class org.apache.kafka.streams.processor.TaskId
-