- AbstractConfig - Class in org.apache.kafka.common.config
-
A convenient base class for configurations to extend.
- AbstractConfig(ConfigDef, Map<?, ?>, boolean) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractConfig(ConfigDef, Map<?, ?>) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractConfig(Map<String, Object>) - Constructor for class org.apache.kafka.common.config.AbstractConfig
-
- AbstractProcessor<K,V> - Class in org.apache.kafka.streams.processor
-
- AbstractProcessor() - Constructor for class org.apache.kafka.streams.processor.AbstractProcessor
-
- accepts(StateStore) - Method in interface org.apache.kafka.streams.state.QueryableStoreType
-
Called when searching for
StateStore
s to see if they
match the type expected by implementors of this interface
- ACKS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
acks
- addClientSaslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SaslConfigs
-
- addClientSslSupport(ConfigDef) - Static method in class org.apache.kafka.common.config.SslConfigs
-
- addDeserializerToConfig(Map<String, Object>, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- addDeserializerToConfig(Properties, Deserializer<?>, Deserializer<?>) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- addErrorMessage(String) - Method in class org.apache.kafka.common.config.ConfigValue
-
- addGlobalStore(StateStore, String, Deserializer, Deserializer, String, String, ProcessorSupplier) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- addInternalTopic(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds an internal topic
- addLatencyAndThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Add a latency and throughput sensor for a specific operation, which will include the following sensors:
average latency
max latency
throughput (num.operations / time unit)
Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the
same scope if it has not been created.
- addProcessor(String, ProcessorSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new processor node that receives and processes records output by one or more parent source or processor node.
- addRecord(ConsumerRecord<K, V>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- addSensor(String, Sensor.RecordingLevel) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Generic method to create a sensor.
- addSensor(String, Sensor.RecordingLevel, Sensor...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Generic method to create a sensor with parent sensors.
- addSerializerToConfig(Map<String, Object>, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- addSerializerToConfig(Properties, Serializer<?>, Serializer<?>) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- addSink(String, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, StreamPartitioner, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic, using
the supplied partitioner.
- addSink(String, String, Serializer, Serializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSink(String, String, Serializer<K>, Serializer<V>, StreamPartitioner<? super K, ? super V>, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new sink that forwards records from upstream parent processor and/or source nodes to the named Kafka topic.
- addSource(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forward the records to child processor and/or sink nodes.
- addSource(String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forward the records to child processor and/or sink nodes.
- addSource(String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, Deserializer, Deserializer, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes the named topics and forwards the records to child processor and/or sink nodes.
- addSource(String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forwards the records to child processor and/or sink nodes.
- addSource(TopologyBuilder.AutoOffsetReset, String, Deserializer, Deserializer, Pattern) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Add a new source that consumes from topics matching the given pattern
and forwards the records to child processor and/or sink nodes.
- addStateStore(StateStoreSupplier, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Adds a state store
- addThroughputSensor(String, String, String, Sensor.RecordingLevel, String...) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Add a throughput sensor for a specific operation:
throughput (num.operations / time unit)
Also create a parent sensor with the same metrics that aggregates all entities with the same operation under the
same scope if it has not been created.
- advanceBy(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return a window definition with the original size, but advance ("hop") the window by the given interval, which
specifies by how much a window moves forward relative to the previous one.
- advanceMs - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the window's advance interval in milliseconds, i.e., by how much a window moves forward relative to
the previous one.
- after(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Changes the end window boundary to timeDifferenceMs
but keep the start window boundary as is.
- afterMs - Variable in class org.apache.kafka.streams.kstream.JoinWindows
-
Maximum time difference for tuples that are after the join tuple.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined windows.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined windows.
- aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined
SessionWindows
.
- aggregate(Initializer<T>, Aggregator<? super K, ? super V, T>, Merger<? super K, T>, SessionWindows, Serde<T>, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Aggregate the values of records in this stream by the grouped key and defined
SessionWindows
.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, Serde<VR>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- aggregate(Initializer<VR>, Aggregator<? super K, ? super V, VR>, Aggregator<? super K, ? super V, VR>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Aggregate the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
using default serializers and deserializers.
- Aggregator<K,V,VA> - Interface in org.apache.kafka.streams.kstream
-
The Aggregator
interface for aggregating values of the given key.
- all() - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Return an iterator over all keys in this store.
- allMetadata() - Method in class org.apache.kafka.streams.KafkaStreams
-
Find all currently running
KafkaStreams
instances (potentially remotely) that use the same
application ID
as this instance (i.e., all instances that belong to
the same Kafka Streams application) and return
StreamsMetadata
for each discovered instance.
- allMetadataForStore(String) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find all currently running
KafkaStreams
instances (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and return
StreamsMetadata
for each discovered instance.
- allPartitionsSorted(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- AlreadyExistsException - Exception in org.apache.kafka.connect.errors
-
Indicates the operation tried to create an entity that already exists.
- AlreadyExistsException(String) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- AlreadyExistsException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- AlreadyExistsException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.AlreadyExistsException
-
- ApiException - Exception in org.apache.kafka.common.errors
-
Any API exception that is part of the public protocol and should be a subclass of this class and be part of this
package.
- ApiException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException(String) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException(Throwable) - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- ApiException() - Constructor for exception org.apache.kafka.common.errors.ApiException
-
- appConfigs() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns all the application config properties as key/value pairs.
- appConfigsWithPrefix(String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns all the application config properties with the given key prefix, as key/value pairs
stripping the prefix.
- APPLICATION_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
application.id
- APPLICATION_SERVER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
user.endpoint
- applicationId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the application id
- apply(R) - Method in interface org.apache.kafka.connect.transforms.Transformation
-
Apply transformation to the record
and return another record object (which may be record
itself) or null
,
corresponding to a map or filter operation respectively.
- apply(K, V, VA) - Method in interface org.apache.kafka.streams.kstream.Aggregator
-
Compute a new aggregate from the key and value of a record and the current aggregate of the same key.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.ForeachAction
-
Perform an action for each record of a stream.
- apply() - Method in interface org.apache.kafka.streams.kstream.Initializer
-
Return the initial value for an aggregation.
- apply(K, V) - Method in interface org.apache.kafka.streams.kstream.KeyValueMapper
-
Map a record with the given key and value to a new value.
- apply(K, V, V) - Method in interface org.apache.kafka.streams.kstream.Merger
-
Compute a new aggregate from the key and two aggregates.
- apply(V, V) - Method in interface org.apache.kafka.streams.kstream.Reducer
-
Aggregate the two given values into a single one.
- apply(V1, V2) - Method in interface org.apache.kafka.streams.kstream.ValueJoiner
-
Return a joined value consisting of value1
and value2
.
- apply(V) - Method in interface org.apache.kafka.streams.kstream.ValueMapper
-
Map the given value to a new value.
- approximateNumEntries() - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Return an approximate count of key-value mappings in this store.
- array(Schema) - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- assign(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- assign(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Manually assign a list of partitions to this consumer.
- assign(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- assign(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RangeAssignor
-
- assign(Map<String, Integer>, Map<String, List<String>>) - Method in class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- assignment() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- assignment() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the set of partitions currently assigned to this consumer.
- assignment() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- assignment() - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Get the current set of assigned TopicPartitions for this task.
- atLeast(Number) - Static method in class org.apache.kafka.common.config.ConfigDef.Range
-
A numeric range that checks only the lower bound
- AuthenticationException - Exception in org.apache.kafka.common.errors
-
- AuthenticationException(String) - Constructor for exception org.apache.kafka.common.errors.AuthenticationException
-
- AuthenticationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.AuthenticationException
-
- AuthorizationException - Exception in org.apache.kafka.common.errors
-
- AuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.AuthorizationException
-
- AuthorizationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.AuthorizationException
-
- AUTO_COMMIT_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
auto.commit.interval.ms
- AUTO_OFFSET_RESET_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
auto.offset.reset
- AUTO_OFFSET_RESET_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- availablePartitionsForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the list of available partitions for this topic
- BATCH_SIZE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
batch.size
- before(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Changes the start window boundary to timeDifferenceMs
but keep the end window boundary as is.
- beforeMs - Variable in class org.apache.kafka.streams.kstream.JoinWindows
-
Maximum time difference for tuples that are before the join tuple.
- beginningOffsets(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- beginningOffsets(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the first offset for the given partitions.
- beginningOffsets(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- between(Number, Number) - Static method in class org.apache.kafka.common.config.ConfigDef.Range
-
A numeric range that checks both the upper and lower bound
- BLOCK_ON_BUFFER_FULL_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- bool() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- BOOLEAN_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- bootstrap(List<InetSocketAddress>) - Static method in class org.apache.kafka.common.Cluster
-
Create a "bootstrap" cluster using the given list of host/ports
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
bootstrap.servers
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
bootstrap.servers
- BOOTSTRAP_SERVERS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
bootstrap.servers
- branch(Predicate<? super K, ? super V>...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Creates an array of KStream
from this stream by branching the records in the original stream based on
the supplied predicates.
- BrokerNotAvailableException - Exception in org.apache.kafka.common.errors
-
- BrokerNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.BrokerNotAvailableException
-
- BrokerNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.BrokerNotAvailableException
-
- BUFFER_MEMORY_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
buffer.memory
- BUFFERED_RECORDS_PER_PARTITION_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
buffered.records.per.partition
- BufferExhaustedException - Exception in org.apache.kafka.clients.producer
-
This exception is thrown if the producer is in non-blocking mode and the rate of data production exceeds the rate at
which data can be sent for long enough for the alloted buffer to be exhausted.
- BufferExhaustedException(String) - Constructor for exception org.apache.kafka.clients.producer.BufferExhaustedException
-
- build() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Build the Schema using the current settings
- build(Integer) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Build the topology for the specified topic group.
- build() - Method in interface org.apache.kafka.streams.state.Stores.InMemoryKeyValueFactory
-
Return the instance of StateStoreSupplier of new key-value store.
- build() - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Return the instance of StateStoreSupplier of new key-value store.
- builder() - Static method in class org.apache.kafka.connect.data.Date
-
Returns a SchemaBuilder for a Date.
- builder(int) - Static method in class org.apache.kafka.connect.data.Decimal
-
Returns a SchemaBuilder for a Decimal with the given scale factor.
- builder() - Static method in class org.apache.kafka.connect.data.Time
-
Returns a SchemaBuilder for a Time.
- builder() - Static method in class org.apache.kafka.connect.data.Timestamp
-
Returns a SchemaBuilder for a Timestamp.
- buildGlobalStateTopology() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Builds the topology for any global state stores
- ByteArray() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- ByteArrayDeserializer - Class in org.apache.kafka.common.serialization
-
- ByteArrayDeserializer() - Constructor for class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- ByteArraySerializer - Class in org.apache.kafka.common.serialization
-
- ByteArraySerializer() - Constructor for class org.apache.kafka.common.serialization.ByteArraySerializer
-
- ByteBuffer() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- ByteBufferDeserializer - Class in org.apache.kafka.common.serialization
-
- ByteBufferDeserializer() - Constructor for class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- ByteBufferSerializer - Class in org.apache.kafka.common.serialization
-
- ByteBufferSerializer() - Constructor for class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- Bytes() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- bytes() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- BYTES_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- BytesDeserializer - Class in org.apache.kafka.common.serialization
-
- BytesDeserializer() - Constructor for class org.apache.kafka.common.serialization.BytesDeserializer
-
- BytesSerializer - Class in org.apache.kafka.common.serialization
-
- BytesSerializer() - Constructor for class org.apache.kafka.common.serialization.BytesSerializer
-
- CACHE_MAX_BYTES_BUFFERING_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
cache.max.bytes.buffering
- Callback - Interface in org.apache.kafka.clients.producer
-
A callback interface that the user can implement to allow code to execute when the request is complete.
- CHECK_CRCS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
check.crcs
- checksum() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The checksum (CRC32) of the record.
- checksum() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The checksum (CRC32) of the record.
- cleanUp() - Method in class org.apache.kafka.streams.KafkaStreams
-
- clear() - Method in class org.apache.kafka.clients.producer.MockProducer
-
Clear the stored history of sent records
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
client.id
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
client.id
- CLIENT_ID_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
client.id
- close() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- close(long, TimeUnit) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- close() - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
This is called when interceptor is closed
- close() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Close the consumer, waiting for up to the default timeout of 30 seconds for any needed cleanup.
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Tries to close the consumer cleanly within the specified timeout.
- close() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- close() - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Close this producer.
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
This method waits up to timeout
for the producer to complete the sending of all incomplete requests.
- close() - Method in class org.apache.kafka.clients.producer.MockProducer
-
- close(long, TimeUnit) - Method in class org.apache.kafka.clients.producer.MockProducer
-
- close() - Method in interface org.apache.kafka.clients.producer.Partitioner
-
This is called when partitioner is closed.
- close() - Method in interface org.apache.kafka.clients.producer.Producer
-
Close this producer
- close(long, TimeUnit) - Method in interface org.apache.kafka.clients.producer.Producer
-
Tries to close the producer cleanly within the specified timeout.
- close() - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
This is called when interceptor is closed
- close() - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- close() - Method in interface org.apache.kafka.common.serialization.Deserializer
-
- close() - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- close() - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- close() - Method in interface org.apache.kafka.common.serialization.Serde
-
Close this serde class, which will close the underlying serializer and deserializer.
- close() - Method in class org.apache.kafka.common.serialization.Serdes.WrapperSerde
-
- close() - Method in interface org.apache.kafka.common.serialization.Serializer
-
Close this serializer.
- close() - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- close() - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- close(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
The SinkTask use this method to close writers for partitions that are no
longer assigned to the SinkTask.
- close() - Method in interface org.apache.kafka.connect.transforms.Transformation
-
Signal that this transformation instance will no longer will be used.
- close() - Method in class org.apache.kafka.streams.KafkaStreams
-
Shutdown this KafkaStreams
instance by signaling all the threads to stop, and then wait for them to join.
- close(long, TimeUnit) - Method in class org.apache.kafka.streams.KafkaStreams
-
Shutdown this KafkaStreams
by signaling all the threads to stop, and then wait up to the timeout for the
threads to join.
- close() - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Close this processor and clean up any resources.
- close() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.Processor
-
Close this processor and clean up any resources.
- close() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Close the storage engine.
- close() - Method in interface org.apache.kafka.streams.state.KeyValueIterator
-
- close() - Method in interface org.apache.kafka.streams.state.WindowStoreIterator
-
- closed() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- Cluster - Class in org.apache.kafka.common
-
A representation of a subset of the nodes, topics, and partitions in the Kafka cluster.
- Cluster(Collection<Node>, Collection<PartitionInfo>, Set<String>) - Constructor for class org.apache.kafka.common.Cluster
-
Deprecated.
Use the Cluster constructor with 5 parameters
- Cluster(String, Collection<Node>, Collection<PartitionInfo>, Set<String>, Set<String>) - Constructor for class org.apache.kafka.common.Cluster
-
Create a new cluster with the given id, nodes and partitions
- ClusterAuthorizationException - Exception in org.apache.kafka.common.errors
-
- ClusterAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.ClusterAuthorizationException
-
- ClusterAuthorizationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ClusterAuthorizationException
-
- clusterId() - Method in class org.apache.kafka.common.ClusterResource
-
Return the cluster id.
- clusterResource() - Method in class org.apache.kafka.common.Cluster
-
- ClusterResource - Class in org.apache.kafka.common
-
The ClusterResource
class encapsulates metadata for a Kafka cluster.
- ClusterResource(String) - Constructor for class org.apache.kafka.common.ClusterResource
-
- ClusterResourceListener - Interface in org.apache.kafka.common
-
A callback interface that users can implement when they wish to get notified about changes in the Cluster metadata.
- commit() - Method in class org.apache.kafka.connect.source.SourceTask
-
- commit() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Requests a commit
- COMMIT_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
commit.interval.ms
- commitAsync() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync(OffsetCommitCallback) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitAsync() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for all the subscribed list of topics and partition.
- commitAsync(OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for the subscribed list of topics and partitions.
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit the specified offsets for the specified list of topics and partitions to Kafka.
- commitAsync(Map<TopicPartition, OffsetAndMetadata>, OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitAsync() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitAsync(OffsetCommitCallback) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- CommitFailedException - Exception in org.apache.kafka.clients.consumer
-
- CommitFailedException() - Constructor for exception org.apache.kafka.clients.consumer.CommitFailedException
-
- commitRecord(SourceRecord) - Method in class org.apache.kafka.connect.source.SourceTask
-
Commit an individual
SourceRecord
when the callback from the producer client is received, or if a record is filtered by a transformation.
- commitSync() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- commitSync() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit offsets returned on the last
poll()
for all the subscribed list of topics and partitions.
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Commit the specified offsets for the specified list of topics and partitions.
- commitSync(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- commitSync() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- committed(TopicPartition) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- committed(TopicPartition) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the last committed offset for the given partition (whether the commit happened by this process or
another).
- committed(TopicPartition) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- compareTo(TaskId) - Method in class org.apache.kafka.streams.processor.TaskId
-
- completeNext() - Method in class org.apache.kafka.clients.producer.MockProducer
-
Complete the earliest uncompleted call successfully.
- COMPRESSION_TYPE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
compression.type
- Config - Class in org.apache.kafka.common.config
-
- Config(List<ConfigValue>) - Constructor for class org.apache.kafka.common.config.Config
-
- config() - Method in class org.apache.kafka.connect.connector.Connector
-
Define the configuration for the connector.
- config() - Method in interface org.apache.kafka.connect.transforms.Transformation
-
Configuration specification for this transformation.
- ConfigDef - Class in org.apache.kafka.common.config
-
This class is used for specifying the set of expected configurations.
- ConfigDef() - Constructor for class org.apache.kafka.common.config.ConfigDef
-
- ConfigDef(ConfigDef) - Constructor for class org.apache.kafka.common.config.ConfigDef
-
- configDef() - Static method in class org.apache.kafka.streams.StreamsConfig
-
Return a copy of the config definition.
- ConfigDef.ConfigKey - Class in org.apache.kafka.common.config
-
- ConfigDef.ConfigKey(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Constructor for class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- ConfigDef.Importance - Enum in org.apache.kafka.common.config
-
The importance level for a configuration
- ConfigDef.Range - Class in org.apache.kafka.common.config
-
Validation logic for numeric ranges
- ConfigDef.Recommender - Interface in org.apache.kafka.common.config
-
This is used by the
ConfigDef.validate(Map)
to get valid values for a configuration given the current
configuration values in order to perform full configuration validation and visibility modification.
- ConfigDef.Type - Enum in org.apache.kafka.common.config
-
The config types
- ConfigDef.Validator - Interface in org.apache.kafka.common.config
-
Validation logic the user may provide to perform single configuration validation.
- ConfigDef.ValidList - Class in org.apache.kafka.common.config
-
- ConfigDef.ValidString - Class in org.apache.kafka.common.config
-
- ConfigDef.Width - Enum in org.apache.kafka.common.config
-
The width of a configuration value
- ConfigException - Exception in org.apache.kafka.common.config
-
Thrown if the user supplies an invalid configuration
- ConfigException(String) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- ConfigException(String, Object) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- ConfigException(String, Object, String) - Constructor for exception org.apache.kafka.common.config.ConfigException
-
- configKeys() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the configuration keys
- configNames() - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- configNames() - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- Configurable - Interface in org.apache.kafka.common
-
A Mix-in style interface for classes that are instantiated by reflection and need to take configuration parameters
- configure(Map<String, ?>) - Method in interface org.apache.kafka.common.Configurable
-
Configure this class with the given key-value pairs
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Deserializer
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Serde
-
Configure this class, which will configure the underlying serializer and deserializer.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.Serdes.WrapperSerde
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.common.serialization.Serializer
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- configure(Map<String, ?>, boolean) - Method in interface org.apache.kafka.connect.storage.Converter
-
Configure this class.
- configure(Map<String, ?>, boolean) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- ConfigValue - Class in org.apache.kafka.common.config
-
- ConfigValue(String) - Constructor for class org.apache.kafka.common.config.ConfigValue
-
- ConfigValue(String, Object, List<Object>, List<String>) - Constructor for class org.apache.kafka.common.config.ConfigValue
-
- configValues() - Method in class org.apache.kafka.common.config.Config
-
- ConnectException - Exception in org.apache.kafka.connect.errors
-
ConnectException is the top-level exception type generated by Kafka Connect and connector implementations.
- ConnectException(String) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- ConnectException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- ConnectException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.ConnectException
-
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
connections.max.idle.ms
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
connections.max.idle.ms
- CONNECTIONS_MAX_IDLE_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- Connector - Class in org.apache.kafka.connect.connector
-
Connectors manage integration of Kafka Connect with another system, either as an input that ingests
data into Kafka or an output that passes data to an external system.
- Connector() - Constructor for class org.apache.kafka.connect.connector.Connector
-
- ConnectorContext - Interface in org.apache.kafka.connect.connector
-
ConnectorContext allows Connectors to proactively interact with the Kafka Connect runtime.
- ConnectorUtils - Class in org.apache.kafka.connect.util
-
Utilities that connector implementations might find useful.
- ConnectorUtils() - Constructor for class org.apache.kafka.connect.util.ConnectorUtils
-
- connectProcessorAndStateStores(String, String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects the processor and the state stores
- connectProcessors(String...) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Connects a list of processors.
- ConnectRecord<R extends ConnectRecord<R>> - Class in org.apache.kafka.connect.connector
-
Base class for records containing data to be copied to/from Kafka.
- ConnectRecord(String, Integer, Schema, Object, Schema, Object, Long) - Constructor for class org.apache.kafka.connect.connector.ConnectRecord
-
- ConnectSchema - Class in org.apache.kafka.connect.data
-
- ConnectSchema(Schema.Type, boolean, Object, String, Integer, String, Map<String, String>, List<Field>, Schema, Schema) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a Schema.
- ConnectSchema(Schema.Type, boolean, Object, String, Integer, String) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a Schema for a primitive type, setting schema parameters, struct fields, and key and value schemas to null.
- ConnectSchema(Schema.Type) - Constructor for class org.apache.kafka.connect.data.ConnectSchema
-
Construct a default schema for a primitive type.
- connectSourceStoreAndTopic(String, String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
This is used only for KStreamBuilder: when adding a KTable from a source topic,
we need to add the topic as the KTable's materialized state store's changelog.
- Consumer<K,V> - Interface in org.apache.kafka.clients.consumer
-
- CONSUMER_PREFIX - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- ConsumerConfig - Class in org.apache.kafka.clients.consumer
-
The consumer configuration keys
- ConsumerInterceptor<K,V> - Interface in org.apache.kafka.clients.consumer
-
A plugin interface that allows you to intercept (and possibly mutate) records received by the consumer.
- consumerPrefix(String) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- ConsumerRebalanceListener - Interface in org.apache.kafka.clients.consumer
-
A callback interface that the user can implement to trigger custom actions when the set of partitions assigned to the
consumer changes.
- ConsumerRecord<K,V> - Class in org.apache.kafka.clients.consumer
-
A key/value pair to be received from Kafka.
- ConsumerRecord(String, int, long, K, V) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecord
-
Creates a record to be received from a specified topic and partition (provided for
compatibility with Kafka 0.9 before the message format supported timestamps and before
serialized metadata were exposed).
- ConsumerRecord(String, int, long, long, TimestampType, long, int, int, K, V) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecord
-
Creates a record to be received from a specified topic and partition
- ConsumerRecords<K,V> - Class in org.apache.kafka.clients.consumer
-
A container that holds the list
ConsumerRecord
per partition for a
particular topic.
- ConsumerRecords(Map<TopicPartition, List<ConsumerRecord<K, V>>>) - Constructor for class org.apache.kafka.clients.consumer.ConsumerRecords
-
- context - Variable in class org.apache.kafka.connect.connector.Connector
-
- context - Variable in class org.apache.kafka.connect.sink.SinkTask
-
- context - Variable in class org.apache.kafka.connect.source.SourceTask
-
- context() - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- ControllerMovedException - Exception in org.apache.kafka.common.errors
-
- ControllerMovedException(String) - Constructor for exception org.apache.kafka.common.errors.ControllerMovedException
-
- ControllerMovedException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ControllerMovedException
-
- Converter - Interface in org.apache.kafka.connect.storage
-
The Converter interface provides support for translating between Kafka Connect's runtime data format
and byte[].
- convertToString(Object, ConfigDef.Type) - Static method in class org.apache.kafka.common.config.ConfigDef
-
- copartitionGroups() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the copartition groups.
- copartitionSources(Collection<String>) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Asserts that the streams of the specified source nodes must be copartitioned.
- CorruptRecordException - Exception in org.apache.kafka.common.errors
-
This exception indicates a record has failed its internal CRC check, this generally indicates network or disk
corruption.
- CorruptRecordException() - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(String) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(Throwable) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- CorruptRecordException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.CorruptRecordException
-
- count() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
The number of records for all topics
- count(String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key.
- count(StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key.
- count(Windows<W>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key and the defined windows.
- count(Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key and the defined windows.
- count(SessionWindows, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key into
SessionWindows
.
- count(SessionWindows, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Count the number of records in this stream by the grouped key into
SessionWindows
.
- count(String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of the original
KTable
that got
mapped
to
the same key into a new instance of
KTable
.
- count(StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Count number of records of the original
KTable
that got
mapped
to
the same key into a new instance of
KTable
.
- create(StateStoreProvider, String) - Method in interface org.apache.kafka.streams.state.QueryableStoreType
-
Create an instance of T (usually a facade) that developers can use
to query the underlying
StateStore
s
- create(String) - Static method in class org.apache.kafka.streams.state.Stores
-
- DataException - Exception in org.apache.kafka.connect.errors
-
Base class for all Kafka Connect data API exceptions.
- DataException(String) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- DataException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- DataException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.DataException
-
- Date - Class in org.apache.kafka.connect.data
-
A date representing a calendar day with no time of day or timezone.
- Date() - Constructor for class org.apache.kafka.connect.data.Date
-
- Decimal - Class in org.apache.kafka.connect.data
-
An arbitrary-precision signed decimal number.
- Decimal() - Constructor for class org.apache.kafka.connect.data.Decimal
-
- DEFAULT_EXCLUDE_INTERNAL_TOPICS - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- DEFAULT_FETCH_MAX_BYTES - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- DEFAULT_KERBEROS_KINIT_CMD - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_MIN_TIME_BEFORE_RELOGIN - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_TICKET_RENEW_JITTER - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_KERBEROS_TICKET_RENEW_WINDOW_FACTOR - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_MAX_PARTITION_FETCH_BYTES - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- DEFAULT_PRINCIPAL_BUILDER_CLASS - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SASL_ENABLED_MECHANISMS - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SASL_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- DEFAULT_SECURITY_PROTOCOL - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- DEFAULT_SSL_ENABLED_PROTOCOLS - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_KEYMANGER_ALGORITHM - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_KEYSTORE_TYPE - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_PROTOCOL - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_TRUSTMANAGER_ALGORITHM - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DEFAULT_SSL_TRUSTSTORE_TYPE - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- DefaultPartitionGrouper - Class in org.apache.kafka.streams.processor
-
Default implementation of the
PartitionGrouper
interface that groups partitions by the partition id.
- DefaultPartitionGrouper() - Constructor for class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
- defaultValue - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- defaultValue() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- defaultValue() - Method in interface org.apache.kafka.connect.data.Schema
-
- defaultValue() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- defaultValue(Object) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the default value for this schema.
- define(ConfigDef.ConfigKey) - Method in class org.apache.kafka.common.config.ConfigDef
-
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no dependents
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic, not dependents and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value and no special validation logic
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, List<String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String, ConfigDef.Recommender) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic and no custom recommender
- define(String, ConfigDef.Type, ConfigDef.Importance, String, String, int, ConfigDef.Width, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value, no special validation logic, no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Validator, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no group, no order in group, no width, no display name, no dependents and no custom recommender
- define(String, ConfigDef.Type, Object, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no special validation logic
- define(String, ConfigDef.Type, ConfigDef.Importance, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
Define a new configuration with no default value and no special validation logic
- delete(K) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Delete the value from the store (if there is one)
- dependents - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- description() - Method in class org.apache.kafka.common.MetricName
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteArrayDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteBufferDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.BytesDeserializer
-
- deserialize(String, byte[]) - Method in interface org.apache.kafka.common.serialization.Deserializer
-
Deserialize a record value from a bytearray into a value or object.
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.DoubleDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.IntegerDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.LongDeserializer
-
- deserialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.StringDeserializer
-
- Deserializer<T> - Interface in org.apache.kafka.common.serialization
-
- deserializer() - Method in interface org.apache.kafka.common.serialization.Serde
-
- deserializer() - Method in class org.apache.kafka.common.serialization.Serdes.WrapperSerde
-
- disableLogging() - Method in interface org.apache.kafka.streams.state.Stores.InMemoryKeyValueFactory
-
Indicates that a changelog should not be created for the key-value store
- disableLogging() - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Indicates that a changelog should not be created for the key-value store
- DisconnectException - Exception in org.apache.kafka.common.errors
-
Server disconnected before a request could be completed.
- DisconnectException() - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(String) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- DisconnectException(Throwable) - Constructor for exception org.apache.kafka.common.errors.DisconnectException
-
- displayName - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- doc() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- doc() - Method in interface org.apache.kafka.connect.data.Schema
-
- doc() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- doc(String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the documentation for this schema.
- documentation - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- Double() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- DoubleDeserializer - Class in org.apache.kafka.common.serialization
-
- DoubleDeserializer() - Constructor for class org.apache.kafka.common.serialization.DoubleDeserializer
-
- DoubleSerializer - Class in org.apache.kafka.common.serialization
-
- DoubleSerializer() - Constructor for class org.apache.kafka.common.serialization.DoubleSerializer
-
- FailOnInvalidTimestamp - Class in org.apache.kafka.streams.processor
-
Retrieves embedded metadata timestamps from Kafka messages.
- FailOnInvalidTimestamp() - Constructor for class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
-
- fetch(K) - Method in interface org.apache.kafka.streams.state.ReadOnlySessionStore
-
Retrieve all aggregated sessions for the provided key
- fetch(K, long, long) - Method in interface org.apache.kafka.streams.state.ReadOnlyWindowStore
-
Get all the key-value pairs with the given key and the time range from all
the existing windows.
- FETCH_MAX_BYTES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
fetch.max.bytes
- FETCH_MAX_WAIT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
fetch.max.wait.ms
- FETCH_MIN_BYTES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
fetch.min.bytes
- field(String) - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- Field - Class in org.apache.kafka.connect.data
-
A field in a
Struct
, consisting of a field name, index, and
Schema
for the field value.
- Field(String, int, Schema) - Constructor for class org.apache.kafka.connect.data.Field
-
- field(String) - Method in interface org.apache.kafka.connect.data.Schema
-
Get a field for this Schema by name.
- field(String, Schema) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Add a field to this struct schema.
- field(String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- fields() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- fields() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the list of fields for this Schema.
- fields() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Get the list of fields for this Schema.
- fillInStackTrace() - Method in exception org.apache.kafka.common.errors.ApiException
-
- fillInStackTrace() - Method in exception org.apache.kafka.common.errors.SerializationException
-
- filter(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
that consists of all records of this stream which satisfy the given predicate.
- filter(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists of all records of this KTable
which satisfy the given
predicate.
- filterNot(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
that consists all records of this stream which do not satisfy the given
predicate.
- filterNot(Predicate<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
that consists all records of this KTable
which do not satisfy the
given predicate.
- findSessions(K, long, long) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Fetch any sessions with the matching key and the sessions end is &le earliestEndTime and the sessions
start is &ge latestStartTime
- flatMap(KeyValueMapper<? super K, ? super V, ? extends Iterable<? extends KeyValue<? extends KR, ? extends VR>>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into zero or more records in the output stream (both key and value type
can be altered arbitrarily).
- flatMapValues(ValueMapper<? super V, ? extends Iterable<? extends VR>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Create a new KStream
by transforming the value of each record in this stream into zero or more values
with the same key in the new stream.
- float32() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- FLOAT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- float64() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- FLOAT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- flush() - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Invoking this method makes all buffered records immediately available to send (even if linger.ms
is
greater than 0) and blocks on the completion of the requests associated with these records.
- flush() - Method in class org.apache.kafka.clients.producer.MockProducer
-
- flush() - Method in interface org.apache.kafka.clients.producer.Producer
-
Flush any accumulated records from the producer.
- flush(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- flush() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Flush any cached data
- foreach(ForeachAction<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Perform an action on each record of KStream
.
- foreach(ForeachAction<? super K, ? super V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Perform an action on each update record of this KTable
.
- ForeachAction<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
ForeachAction
interface for performing an action on a
key-value
pair
.
- forward(K, V) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to the downstream processors
- forward(K, V, int) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by childIndex
- forward(K, V, String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Forwards a key/value pair to one of the downstream processors designated by the downstream processor name
- fromConnectData(String, Schema, Object) - Method in interface org.apache.kafka.connect.storage.Converter
-
Convert a Kafka Connect data object to a native object for serialization.
- fromConnectData(String, Schema, Object) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Date
-
Convert a value from its logical format (Date) to it's encoded format.
- fromLogical(Schema, BigDecimal) - Static method in class org.apache.kafka.connect.data.Decimal
-
Convert a value from its logical format (BigDecimal) to it's encoded format.
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Time
-
Convert a value from its logical format (Time) to it's encoded format.
- fromLogical(Schema, Date) - Static method in class org.apache.kafka.connect.data.Timestamp
-
Convert a value from its logical format (Date) to it's encoded format.
- get(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- get(String) - Method in class org.apache.kafka.connect.data.Struct
-
Get the value of a field, returning the default value if no value has been set yet and a default value is specified
in the field's schema.
- get(Field) - Method in class org.apache.kafka.connect.data.Struct
-
Get the value of a field, returning the default value if no value has been set yet and a default value is specified
in the field's schema.
- get() - Method in interface org.apache.kafka.streams.kstream.TransformerSupplier
-
- get() - Method in interface org.apache.kafka.streams.kstream.ValueTransformerSupplier
-
- get() - Method in interface org.apache.kafka.streams.processor.ProcessorSupplier
-
- get() - Method in interface org.apache.kafka.streams.processor.StateStoreSupplier
-
- get(K) - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Get the value corresponding to this key.
- getArray(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getBoolean(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getBoolean(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getBytes(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getClass(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getConfiguredInstance(String, Class<T>) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get a configured instance of the give class specified by the given configuration key.
- getConfiguredInstances(String, Class<T>) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get a list of configured instances of the given class specified by the given configuration key.
- getConfiguredInstances(String, Class<T>, Map<String, Object>) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get a list of configured instances of the given class specified by the given configuration key.
- getConfigValue(ConfigDef.ConfigKey, String) - Method in class org.apache.kafka.common.config.ConfigDef
-
- getConsumer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
Create a
Consumer
which is used to read records of source topics.
- getConsumerConfigs(StreamThread, String, String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getDouble(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getFloat32(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getFloat64(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getInt16(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt32(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt64(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getInt8(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getList(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getLong(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getMap(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getName() - Method in enum org.apache.kafka.connect.data.Schema.Type
-
- getPassword(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getProducer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
Create a
Producer
which is used to write records to sink topics.
- getProducerConfigs(String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getRestoreConsumer(Map<String, Object>) - Method in interface org.apache.kafka.streams.KafkaClientSupplier
-
- getRestoreConsumerConfigs(String) - Method in class org.apache.kafka.streams.StreamsConfig
-
- getShort(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getStateStore(String) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Get the state store given the store name.
- getStoreName() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Get the name of the local state store used for materializing this KTable
.
- getString(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- getString(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getStruct(String) - Method in class org.apache.kafka.connect.data.Struct
-
- getWithoutDefault(String) - Method in class org.apache.kafka.connect.data.Struct
-
Get the underlying raw value for the field without accounting for default values.
- GlobalKTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
GlobalKTable
is an abstraction of a changelog stream from a primary-keyed table.
- globalStateStores() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Get any global
StateStore
s that are part of the
topology
- globalTable(String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
- globalTable(Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
- group - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- group() - Method in class org.apache.kafka.common.MetricName
-
- GROUP_ID_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
group.id
- GroupAuthorizationException - Exception in org.apache.kafka.common.errors
-
- GroupAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.GroupAuthorizationException
-
- groupBy(KeyValueMapper<? super K, ? super V, KR>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Group the records of this
KStream
on a new key that is selected using the provided
KeyValueMapper
and default serializers and deserializers.
- groupBy(KeyValueMapper<? super K, ? super V, KR>, Serde<KR>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Group the records of this
KStream
on a new key that is selected using the provided
KeyValueMapper
.
- groupBy(KeyValueMapper<? super K, ? super V, KeyValue<KR, VR>>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Re-groups the records of this
KTable
using the provided
KeyValueMapper
and default serializers
and deserializers.
- groupBy(KeyValueMapper<? super K, ? super V, KeyValue<KR, VR>>, Serde<KR>, Serde<VR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Re-groups the records of this
KTable
using the provided
KeyValueMapper
.
- groupByKey() - Method in interface org.apache.kafka.streams.kstream.KStream
-
Group the records by their current key into a
KGroupedStream
while preserving the original values
and default serializers and deserializers.
- groupByKey(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Group the records by their current key into a
KGroupedStream
while preserving the original values.
- GroupCoordinatorNotAvailableException - Exception in org.apache.kafka.common.errors
-
The broker returns this error code for consumer metadata requests or offset commit requests if the offsets topic has
not yet been created.
- GroupCoordinatorNotAvailableException() - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- GroupCoordinatorNotAvailableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- groupId() - Method in exception org.apache.kafka.common.errors.GroupAuthorizationException
-
- GroupLoadInProgressException - Exception in org.apache.kafka.common.errors
-
The broker returns this error code for any coordinator request if it is still loading the metadata (after a leader change
for that offsets topic partition) for this group.
- GroupLoadInProgressException() - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(String) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- GroupLoadInProgressException(Throwable) - Constructor for exception org.apache.kafka.common.errors.GroupLoadInProgressException
-
- groupPartitions(List<T>, int) - Static method in class org.apache.kafka.connect.util.ConnectorUtils
-
Given a list of elements and a target number of groups, generates list of groups of
elements to match the target number of groups, spreading them evenly among the groups.
- groups() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the groups for the configuration
- GSSAPI_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- id() - Method in class org.apache.kafka.common.Node
-
The node id of this node
- idString() - Method in class org.apache.kafka.common.Node
-
String representation of the node id.
- ignore(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
- IllegalGenerationException - Exception in org.apache.kafka.common.errors
-
- IllegalGenerationException() - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(String) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalGenerationException(Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalGenerationException
-
- IllegalSaslStateException - Exception in org.apache.kafka.common.errors
-
- IllegalSaslStateException(String) - Constructor for exception org.apache.kafka.common.errors.IllegalSaslStateException
-
- IllegalSaslStateException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.IllegalSaslStateException
-
- IllegalWorkerStateException - Exception in org.apache.kafka.connect.errors
-
Indicates that a method has been invoked illegally or at an invalid time by a connector or task.
- IllegalWorkerStateException(String) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- IllegalWorkerStateException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- IllegalWorkerStateException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.IllegalWorkerStateException
-
- importance - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- in(String...) - Static method in class org.apache.kafka.common.config.ConfigDef.ValidList
-
- in(String...) - Static method in class org.apache.kafka.common.config.ConfigDef.ValidString
-
- inactivityGap() - Method in class org.apache.kafka.streams.kstream.SessionWindows
-
Return the specified gap for the session windows in milliseconds.
- InconsistentGroupProtocolException - Exception in org.apache.kafka.common.errors
-
- InconsistentGroupProtocolException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InconsistentGroupProtocolException
-
- InconsistentGroupProtocolException(String) - Constructor for exception org.apache.kafka.common.errors.InconsistentGroupProtocolException
-
- index() - Method in class org.apache.kafka.connect.data.Field
-
Get the index of this field within the struct.
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Initialize this transformer.
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Initialize this transformer.
- init(ProcessorContext) - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- init(ProcessorContext) - Method in interface org.apache.kafka.streams.processor.Processor
-
Initialize this processor with the given context.
- init(ProcessorContext, StateStore) - Method in interface org.apache.kafka.streams.processor.StateStore
-
Initializes this state store
- initialize(ConnectorContext) - Method in class org.apache.kafka.connect.connector.Connector
-
Initialize this connector, using the provided ConnectorContext to notify the runtime of
input configuration changes.
- initialize(ConnectorContext, List<Map<String, String>>) - Method in class org.apache.kafka.connect.connector.Connector
-
Initialize this connector, using the provided ConnectorContext to notify the runtime of
input configuration changes and using the provided set of Task configurations.
- initialize(SinkTaskContext) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Initialize the context of this task.
- initialize(SourceTaskContext) - Method in class org.apache.kafka.connect.source.SourceTask
-
Initialize this SourceTask with the specified context object.
- Initializer<VA> - Interface in org.apache.kafka.streams.kstream
-
The Initializer
interface for creating an initial value in aggregations.
- inMemory() - Method in interface org.apache.kafka.streams.state.Stores.KeyValueFactory
-
Keep all key-value entries in-memory, although for durability all entries are recorded in a Kafka topic that can be
read to restore the entries if they are lost.
- INSTANCE - Static variable in exception org.apache.kafka.common.errors.DisconnectException
-
- INSTANCE - Static variable in exception org.apache.kafka.common.errors.GroupCoordinatorNotAvailableException
-
- inSyncReplicas() - Method in class org.apache.kafka.common.PartitionInfo
-
The subset of the replicas that are in sync, that is caught-up to the leader and ready to take over as leader if
the leader should fail
- int16() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT16_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int32() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int64() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- int8() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- INT8_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- Integer() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- IntegerDeserializer - Class in org.apache.kafka.common.serialization
-
- IntegerDeserializer() - Constructor for class org.apache.kafka.common.serialization.IntegerDeserializer
-
- IntegerSerializer - Class in org.apache.kafka.common.serialization
-
- IntegerSerializer() - Constructor for class org.apache.kafka.common.serialization.IntegerSerializer
-
- INTERCEPTOR_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
interceptor.classes
- INTERCEPTOR_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
interceptor.classes
- INTERCEPTOR_CLASSES_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- INTERCEPTOR_CLASSES_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- internalTopics() - Method in class org.apache.kafka.common.Cluster
-
- InterruptException - Exception in org.apache.kafka.common.errors
-
An unchecked wrapper for InterruptedException
- InterruptException(InterruptedException) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- InterruptException(String, InterruptedException) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- InterruptException(String) - Constructor for exception org.apache.kafka.common.errors.InterruptException
-
- InvalidCommitOffsetSizeException - Exception in org.apache.kafka.common.errors
-
- InvalidCommitOffsetSizeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidCommitOffsetSizeException
-
- InvalidCommitOffsetSizeException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidCommitOffsetSizeException
-
- InvalidConfigurationException - Exception in org.apache.kafka.common.errors
-
- InvalidConfigurationException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidConfigurationException
-
- InvalidConfigurationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidConfigurationException
-
- InvalidFetchSizeException - Exception in org.apache.kafka.common.errors
-
- InvalidFetchSizeException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidFetchSizeException
-
- InvalidFetchSizeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidFetchSizeException
-
- InvalidGroupIdException - Exception in org.apache.kafka.common.errors
-
- InvalidGroupIdException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidGroupIdException
-
- InvalidGroupIdException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidGroupIdException
-
- InvalidMetadataException - Exception in org.apache.kafka.common.errors
-
An exception that may indicate the client's metadata is out of date
- InvalidMetadataException() - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidMetadataException(Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidMetadataException
-
- InvalidOffsetException - Exception in org.apache.kafka.clients.consumer
-
Thrown when the offset for a set of partitions is invalid (either undefined or out of range),
and no reset policy has been configured.
- InvalidOffsetException(String) - Constructor for exception org.apache.kafka.clients.consumer.InvalidOffsetException
-
- InvalidOffsetException - Exception in org.apache.kafka.common.errors
-
Thrown when the offset for a set of partitions is invalid (either undefined or out of range),
and no reset policy has been configured.
- InvalidOffsetException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidOffsetException
-
- InvalidOffsetException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidOffsetException
-
- InvalidPartitionsException - Exception in org.apache.kafka.common.errors
-
- InvalidPartitionsException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidPartitionsException
-
- InvalidPartitionsException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidPartitionsException
-
- InvalidReplicaAssignmentException - Exception in org.apache.kafka.common.errors
-
- InvalidReplicaAssignmentException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidReplicaAssignmentException
-
- InvalidReplicaAssignmentException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidReplicaAssignmentException
-
- InvalidReplicationFactorException - Exception in org.apache.kafka.common.errors
-
- InvalidReplicationFactorException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidReplicationFactorException
-
- InvalidReplicationFactorException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidReplicationFactorException
-
- InvalidRequestException - Exception in org.apache.kafka.common.errors
-
Thrown when a request breaks basic wire protocol rules.
- InvalidRequestException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidRequestException
-
- InvalidRequestException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidRequestException
-
- InvalidRequiredAcksException - Exception in org.apache.kafka.common.errors
-
- InvalidRequiredAcksException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidRequiredAcksException
-
- InvalidSessionTimeoutException - Exception in org.apache.kafka.common.errors
-
- InvalidSessionTimeoutException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidSessionTimeoutException
-
- InvalidSessionTimeoutException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidSessionTimeoutException
-
- InvalidStateStoreException - Exception in org.apache.kafka.streams.errors
-
Indicates that there was a problem when trying to access a
StateStore
, i.e, the Store is no longer valid because it is
closed or doesn't exist any more due to a rebalance.
- InvalidStateStoreException(String) - Constructor for exception org.apache.kafka.streams.errors.InvalidStateStoreException
-
- InvalidStateStoreException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.InvalidStateStoreException
-
- InvalidStateStoreException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.InvalidStateStoreException
-
- InvalidTimestampException - Exception in org.apache.kafka.common.errors
-
Indicate the timestamp of a record is invalid.
- InvalidTimestampException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidTimestampException
-
- InvalidTimestampException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTimestampException
-
- InvalidTopicException - Exception in org.apache.kafka.common.errors
-
The client has attempted to perform an operation on an invalid topic.
- InvalidTopicException() - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(String) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- InvalidTopicException(Throwable) - Constructor for exception org.apache.kafka.common.errors.InvalidTopicException
-
- isBootstrapConfigured() - Method in class org.apache.kafka.common.Cluster
-
- isCreatedOrRunning() - Method in enum org.apache.kafka.streams.KafkaStreams.State
-
- isEmpty() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
- isEmpty() - Method in class org.apache.kafka.common.Node
-
Check whether this node is empty, which may be the case if noNode() is used as a placeholder
in a response payload with an error.
- isOpen() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Is this store open for reading and writing
- isOptional() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- isOptional() - Method in interface org.apache.kafka.connect.data.Schema
-
- isOptional() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- isPrimitive() - Method in enum org.apache.kafka.connect.data.Schema.Type
-
- isRunning() - Method in enum org.apache.kafka.streams.KafkaStreams.State
-
- isValidTransition(KafkaStreams.State) - Method in enum org.apache.kafka.streams.KafkaStreams.State
-
- iterator() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
- KafkaClientSupplier - Interface in org.apache.kafka.streams
-
KafkaClientSupplier
can be used to provide custom Kafka clients to a
KafkaStreams
instance.
- KafkaConsumer<K,V> - Class in org.apache.kafka.clients.consumer
-
A client that consumes records from a Kafka cluster.
- KafkaConsumer(Map<String, Object>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a set of key-value pairs as configuration.
- KafkaConsumer(Map<String, Object>, Deserializer<K>, Deserializer<V>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a set of key-value pairs as configuration, and a key and a value
Deserializer
.
- KafkaConsumer(Properties) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a
Properties
object as configuration.
- KafkaConsumer(Properties, Deserializer<K>, Deserializer<V>) - Constructor for class org.apache.kafka.clients.consumer.KafkaConsumer
-
A consumer is instantiated by providing a
Properties
object as configuration, and a
key and a value
Deserializer
.
- KafkaException - Exception in org.apache.kafka.common
-
The base class of all other Kafka exceptions
- KafkaException(String, Throwable) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException(String) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException(Throwable) - Constructor for exception org.apache.kafka.common.KafkaException
-
- KafkaException() - Constructor for exception org.apache.kafka.common.KafkaException
-
- kafkaOffset() - Method in class org.apache.kafka.connect.sink.SinkRecord
-
- kafkaPartition() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- KafkaProducer<K,V> - Class in org.apache.kafka.clients.producer
-
A Kafka client that publishes records to the Kafka cluster.
- KafkaProducer(Map<String, Object>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration.
- KafkaProducer(Map<String, Object>, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration, a key and a value
Serializer
.
- KafkaProducer(Properties) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration.
- KafkaProducer(Properties, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.KafkaProducer
-
A producer is instantiated by providing a set of key-value pairs as configuration, a key and a value
Serializer
.
- KafkaStreams - Class in org.apache.kafka.streams
-
A Kafka client that allows for performing continuous computation on input coming from one or more input topics and
sends output to zero, one, or more output topics.
- KafkaStreams(TopologyBuilder, Properties) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams(TopologyBuilder, StreamsConfig) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams(TopologyBuilder, StreamsConfig, KafkaClientSupplier) - Constructor for class org.apache.kafka.streams.KafkaStreams
-
Create a KafkaStreams
instance.
- KafkaStreams.State - Enum in org.apache.kafka.streams
-
Kafka Streams states are the possible state that a Kafka Streams instance can be in.
- KafkaStreams.StateListener - Interface in org.apache.kafka.streams
-
- key() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The key (or null if no key is specified)
- key() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- key() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- key - Variable in class org.apache.kafka.streams.KeyValue
-
The key of the key-value pair.
- key() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the key of the window.
- KEY_DESERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
key.deserializer
- KEY_DESERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- KEY_SERDE_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
key.serde
- KEY_SERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
key.serializer
- KEY_SERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- keyDeserializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key deserializer.
- keyFrom(byte[]) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Deserialize the key from raw bytes.
- keySchema() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- keySchema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- keySchema() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the key schema for this map schema.
- keySchema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- keySerde() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the default key serde
- keySerde() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serde.
- keySerde() - Method in class org.apache.kafka.streams.StreamsConfig
-
- keySerializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the key serializer.
- KeyValue<K,V> - Class in org.apache.kafka.streams
-
A key-value pair defined for a single Kafka Streams record.
- KeyValue(K, V) - Constructor for class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- KeyValueIterator<K,V> - Interface in org.apache.kafka.streams.state
-
- KeyValueMapper<K,V,VR> - Interface in org.apache.kafka.streams.kstream
-
The
KeyValueMapper
interface for mapping a
key-value pair
to a new value of arbitrary type.
- KeyValueStore<K,V> - Interface in org.apache.kafka.streams.state
-
A key-value store that supports put/get/delete and range queries.
- keyValueStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- KGroupedStream<K,V> - Interface in org.apache.kafka.streams.kstream
-
KGroupedStream
is an abstraction of a
grouped record stream of
KeyValue
pairs.
- KGroupedTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KGroupedTable
is an abstraction of a re-grouped changelog stream from a primary-keyed table,
usually on a different grouping key than the original primary key.
- KStream<K,V> - Interface in org.apache.kafka.streams.kstream
-
KStream
is an abstraction of a
record stream of
KeyValue
pairs, i.e., each record is an
independent entity/event in the real world.
- KStreamBuilder - Class in org.apache.kafka.streams.kstream
-
KStreamBuilder
provide the high-level Kafka Streams DSL to specify a Kafka Streams topology.
- KStreamBuilder() - Constructor for class org.apache.kafka.streams.kstream.KStreamBuilder
-
- KTable<K,V> - Interface in org.apache.kafka.streams.kstream
-
KTable
is an abstraction of a changelog stream from a primary-keyed table.
- main(String[]) - Static method in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- main(String[]) - Static method in class org.apache.kafka.clients.producer.ProducerConfig
-
- main(String[]) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- maintainMs() - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.SessionWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return the window maintain duration (retention time) in milliseconds.
- maintainMs() - Method in class org.apache.kafka.streams.kstream.Windows
-
Return the window maintain duration (retention time) in milliseconds.
- map(Schema, Schema) - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- map(KeyValueMapper<? super K, ? super V, ? extends KeyValue<? extends KR, ? extends VR>>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into a new record in the output stream (both key and value type can be
altered arbitrarily).
- mapValues(ValueMapper<? super V, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform the value of each input record into a new value (with possible new type) of the output record.
- mapValues(ValueMapper<? super V, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Create a new KTable
by transforming the value of each record in this KTable
into a new value
(with possible new type)in the new KTable
.
- MAX_BLOCK_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
max.block.ms
- MAX_IN_FLIGHT_REQUESTS_PER_CONNECTION - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
max.in.flight.requests.per.connection
- MAX_PARTITION_FETCH_BYTES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
max.partition.fetch.bytes
- MAX_POLL_INTERVAL_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
max.poll.interval.ms
- MAX_POLL_RECORDS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
max.poll.records
- MAX_REQUEST_SIZE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
max.request.size
- maxEntries(int) - Method in interface org.apache.kafka.streams.state.Stores.InMemoryKeyValueFactory
-
Limits the in-memory key-value store to hold a maximum number of entries.
- maxNumPartitions(Cluster, Set<String>) - Method in class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
- merge(KStream<K, V>...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
- Merger<K,V> - Interface in org.apache.kafka.streams.kstream
-
The interface for merging aggregate values for
SessionWindows
with the given key.
- metadata() - Method in class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
- METADATA_FETCH_TIMEOUT_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- METADATA_MAX_AGE_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
metadata.max.age.ms
- METADATA_MAX_AGE_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
metadata.max.age.ms
- METADATA_MAX_AGE_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- metadataForKey(String, K, Serializer<K>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find the currently running
KafkaStreams
instance (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and the
StateStore
contains the given
key
and return
StreamsMetadata
for it.
- metadataForKey(String, K, StreamPartitioner<? super K, ?>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Find the currently running
KafkaStreams
instance (potentially remotely) that
use the same
application ID
as this instance (i.e., all
instances that belong to the same Kafka Streams application)
and that contain a
StateStore
with the given
storeName
and the
StateStore
contains the given
key
and return
StreamsMetadata
for it.
- Metric - Interface in org.apache.kafka.common
-
A numerical metric tracked for monitoring purposes
- METRIC_REPORTER_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
metric.reporters
- METRIC_REPORTER_CLASSES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
metric.reporters
- METRIC_REPORTER_CLASSES_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metric.reporters
- metricName() - Method in interface org.apache.kafka.common.Metric
-
A name for this metric
- MetricName - Class in org.apache.kafka.common
-
The MetricName
class encapsulates a metric's name, logical group and its related attributes.
- MetricName(String, String, String, Map<String, String>) - Constructor for class org.apache.kafka.common.MetricName
-
Please create MetricName by method Metrics.metricName(String, String, String, Map)
- MetricName(String, String, String, String...) - Constructor for class org.apache.kafka.common.MetricName
-
Deprecated.
This method will be removed in a future release.
Please create MetricName by method Metrics.metricName(String, String, String, String...)
- MetricName(String, String, Map<String, String>) - Constructor for class org.apache.kafka.common.MetricName
-
Deprecated.
This method will be removed in a future release.
Please create MetricName by method Metrics.metricName(String, String, Map)
- MetricName(String, String, String) - Constructor for class org.apache.kafka.common.MetricName
-
Deprecated.
This method will be removed in a future release.
Please create MetricName by method Metrics.metricName(String, String, String)
- MetricName(String, String) - Constructor for class org.apache.kafka.common.MetricName
-
Deprecated.
This method will be removed in a future release.
Please create MetricName by method Metrics.metricName(String, String)
- metrics() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- metrics() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the metrics kept by the consumer
- metrics() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- metrics() - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Get the full set of internal metrics maintained by the producer.
- metrics() - Method in class org.apache.kafka.clients.producer.MockProducer
-
- metrics() - Method in interface org.apache.kafka.clients.producer.Producer
-
Return a map of metrics maintained by the producer
- metrics() - Method in class org.apache.kafka.streams.KafkaStreams
-
Get read-only handle on global metrics registry.
- metrics() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns Metrics instance
- metrics() - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Get read-only handle on global metrics registry.
- METRICS_NUM_SAMPLES_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
metrics.num.samples
- METRICS_NUM_SAMPLES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
metrics.num.samples
- METRICS_NUM_SAMPLES_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.num.samples
- METRICS_RECORDING_LEVEL_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
metrics.log.level
- METRICS_RECORDING_LEVEL_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.record.level
- METRICS_SAMPLE_WINDOW_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
metrics.sample.window.ms
- METRICS_SAMPLE_WINDOW_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
metrics.sample.window.ms
- METRICS_SAMPLE_WINDOW_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
metrics.sample.window.ms
- MockConsumer<K,V> - Class in org.apache.kafka.clients.consumer
-
A mock of the
Consumer
interface you can use for testing code that uses Kafka.
- MockConsumer(OffsetResetStrategy) - Constructor for class org.apache.kafka.clients.consumer.MockConsumer
-
- MockProducer<K,V> - Class in org.apache.kafka.clients.producer
-
A mock of the producer interface you can use for testing code that uses Kafka.
- MockProducer(Cluster, boolean, Partitioner, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.MockProducer
-
Create a mock producer
- MockProducer(boolean, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.MockProducer
-
- MockProducer(boolean, Partitioner, Serializer<K>, Serializer<V>) - Constructor for class org.apache.kafka.clients.producer.MockProducer
-
- of(long) - Static method in class org.apache.kafka.streams.kstream.JoinWindows
-
Specifies that records of the same key are joinable if their timestamps are within timeDifferenceMs
,
i.e., the timestamp of a record from the secondary stream is max timeDifferenceMs
earlier or later than
the timestamp of the record from the primary stream.
- of(long) - Static method in class org.apache.kafka.streams.kstream.TimeWindows
-
Return a window definition with the given window size, and with the advance interval being equal to the window
size.
- of() - Static method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return an unlimited window starting at timestamp zero.
- offset() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The position of this record in the corresponding Kafka partition.
- offset() - Method in class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
- offset() - Method in class org.apache.kafka.clients.consumer.OffsetAndTimestamp
-
- offset() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The offset of the record in the topic/partition.
- offset(Map<TopicPartition, Long>) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Reset the consumer offsets for the given topic partitions.
- offset(TopicPartition, long) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Reset the consumer offsets for the given topic partition.
- offset(Map<String, T>) - Method in interface org.apache.kafka.connect.storage.OffsetStorageReader
-
Get the offset for the specified partition.
- offset() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the offset of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- OffsetAndMetadata - Class in org.apache.kafka.clients.consumer
-
The Kafka offset commit API allows users to provide additional metadata (in the form of a string)
when an offset is committed.
- OffsetAndMetadata(long, String) - Constructor for class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
Construct a new OffsetAndMetadata object for committing through
KafkaConsumer
.
- OffsetAndMetadata(long) - Constructor for class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
Construct a new OffsetAndMetadata object for committing through
KafkaConsumer
.
- OffsetAndTimestamp - Class in org.apache.kafka.clients.consumer
-
A container class for offset and timestamp.
- OffsetAndTimestamp(long, long) - Constructor for class org.apache.kafka.clients.consumer.OffsetAndTimestamp
-
- OffsetCommitCallback - Interface in org.apache.kafka.clients.consumer
-
A callback interface that the user can implement to trigger custom actions when a commit request completes.
- OffsetMetadataTooLarge - Exception in org.apache.kafka.common.errors
-
The client has tried to save its offset with associated metadata larger than the maximum size allowed by the server.
- OffsetMetadataTooLarge() - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(String) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetMetadataTooLarge(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetMetadataTooLarge
-
- OffsetOutOfRangeException - Exception in org.apache.kafka.clients.consumer
-
No reset policy has been defined, and the offsets for these partitions are either larger or smaller
than the range of offsets the server has for the given partition.
- OffsetOutOfRangeException(Map<TopicPartition, Long>) - Constructor for exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- OffsetOutOfRangeException - Exception in org.apache.kafka.common.errors
-
No reset policy has been defined, and the offsets for these partitions are either larger or smaller
than the range of offsets the server has for the given partition.
- OffsetOutOfRangeException(String) - Constructor for exception org.apache.kafka.common.errors.OffsetOutOfRangeException
-
- OffsetOutOfRangeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.OffsetOutOfRangeException
-
- offsetOutOfRangePartitions() - Method in exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- OffsetResetStrategy - Enum in org.apache.kafka.clients.consumer
-
- offsets(Collection<Map<String, T>>) - Method in interface org.apache.kafka.connect.storage.OffsetStorageReader
-
Get a set of offsets for the specified partition identifiers.
- offsetsForTimes(Map<TopicPartition, Long>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- offsetsForTimes(Map<TopicPartition, Long>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Look up the offsets for the given partitions by timestamp.
- offsetsForTimes(Map<TopicPartition, Long>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- offsetStorageReader() - Method in interface org.apache.kafka.connect.source.SourceTaskContext
-
Get the OffsetStorageReader for this SourceTask.
- OffsetStorageReader - Interface in org.apache.kafka.connect.storage
-
OffsetStorageReader provides access to the offset storage used by sources.
- onAcknowledgement(RecordMetadata, Exception) - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
This method is called when the record sent to the server has been acknowledged, or when sending the record fails before
it gets sent to the server.
- onChange(KafkaStreams.State, KafkaStreams.State) - Method in interface org.apache.kafka.streams.KafkaStreams.StateListener
-
Called when state changes.
- onCommit(Map<TopicPartition, OffsetAndMetadata>) - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
This is called when offsets get committed.
- onComplete(Map<TopicPartition, OffsetAndMetadata>, Exception) - Method in interface org.apache.kafka.clients.consumer.OffsetCommitCallback
-
A callback method the user can implement to provide asynchronous handling of commit request completion.
- onCompletion(RecordMetadata, Exception) - Method in interface org.apache.kafka.clients.producer.Callback
-
A callback method the user can implement to provide asynchronous handling of request completion.
- onConsume(ConsumerRecords<K, V>) - Method in interface org.apache.kafka.clients.consumer.ConsumerInterceptor
-
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.FailOnInvalidTimestamp
-
Raises an exception on every call.
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.LogAndSkipOnInvalidTimestamp
-
Writes a log WARN message when the extracted timestamp is invalid (negative) but returns the invalid timestamp as-is,
which ultimately causes the record to be skipped and not to be processed.
- onInvalidTimestamp(ConsumerRecord<Object, Object>, long, long) - Method in class org.apache.kafka.streams.processor.UsePreviousTimeOnInvalidTimestamp
-
Returns the current stream-time as new timestamp for the record.
- onPartitionsAssigned(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.ConsumerRebalanceListener
-
A callback method the user can implement to provide handling of customized offsets on completion of a successful
partition re-assignment.
- onPartitionsAssigned(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- onPartitionsRevoked(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.ConsumerRebalanceListener
-
A callback method the user can implement to provide handling of offset commits to a customized store on the start
of a rebalance operation.
- onPartitionsRevoked(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
- onSend(ProducerRecord<K, V>) - Method in interface org.apache.kafka.clients.producer.ProducerInterceptor
-
- onUpdate(ClusterResource) - Method in interface org.apache.kafka.common.ClusterResourceListener
-
A callback method that a user can implement to get updates for
ClusterResource
.
- open(Collection<TopicPartition>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
The SinkTask use this method to create writers for newly assigned partitions in case of partition
rebalance.
- optional() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set this schema as optional.
- OPTIONAL_BOOLEAN_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_BYTES_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_FLOAT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_FLOAT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT16_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT32_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT64_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_INT8_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- OPTIONAL_STRING_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- orderInGroup - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- org.apache.kafka.clients.consumer - package org.apache.kafka.clients.consumer
-
- org.apache.kafka.clients.producer - package org.apache.kafka.clients.producer
-
- org.apache.kafka.common - package org.apache.kafka.common
-
- org.apache.kafka.common.config - package org.apache.kafka.common.config
-
- org.apache.kafka.common.errors - package org.apache.kafka.common.errors
-
- org.apache.kafka.common.serialization - package org.apache.kafka.common.serialization
-
- org.apache.kafka.connect.connector - package org.apache.kafka.connect.connector
-
- org.apache.kafka.connect.data - package org.apache.kafka.connect.data
-
- org.apache.kafka.connect.errors - package org.apache.kafka.connect.errors
-
- org.apache.kafka.connect.sink - package org.apache.kafka.connect.sink
-
- org.apache.kafka.connect.source - package org.apache.kafka.connect.source
-
- org.apache.kafka.connect.storage - package org.apache.kafka.connect.storage
-
- org.apache.kafka.connect.transforms - package org.apache.kafka.connect.transforms
-
- org.apache.kafka.connect.util - package org.apache.kafka.connect.util
-
- org.apache.kafka.streams - package org.apache.kafka.streams
-
- org.apache.kafka.streams.errors - package org.apache.kafka.streams.errors
-
- org.apache.kafka.streams.kstream - package org.apache.kafka.streams.kstream
-
- org.apache.kafka.streams.processor - package org.apache.kafka.streams.processor
-
- org.apache.kafka.streams.state - package org.apache.kafka.streams.state
-
- originals() - Method in class org.apache.kafka.common.config.AbstractConfig
-
- originalsStrings() - Method in class org.apache.kafka.common.config.AbstractConfig
-
Get all the original settings, ensuring that all values are of type String.
- originalsWithPrefix(String) - Method in class org.apache.kafka.common.config.AbstractConfig
-
Gets all original settings with the given prefix, stripping the prefix before adding it to the output.
- outerJoin(KStream<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, JoinWindows) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Join records of this stream with another KStream
's records using windowed left equi join with default
serializers and deserializers.
- outerJoin(KStream<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>, JoinWindows, Serde<K>, Serde<V>, Serde<VO>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Join records of this stream with another KStream
's records using windowed left equi join.
- outerJoin(KTable<K, VO>, ValueJoiner<? super V, ? super VO, ? extends VR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Join records of this KTable
(left input) with another KTable
's (right input) records using
non-windowed outer equi join.
- overlap(Window) - Method in class org.apache.kafka.streams.kstream.Window
-
Check if the given window overlaps with this window.
- pair(K, V) - Static method in class org.apache.kafka.streams.KeyValue
-
Create a new key-value pair.
- parameter(String, String) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set a schema parameter.
- parameters() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- parameters() - Method in interface org.apache.kafka.connect.data.Schema
-
Get a map of schema parameters.
- parameters() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- parameters(Map<String, String>) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set schema parameters.
- parse(Map<?, ?>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Parse and validate configs against this configuration definition.
- parse(String) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- parseType(String, Object, ConfigDef.Type) - Static method in class org.apache.kafka.common.config.ConfigDef
-
Parse a value according to its expected type.
- partition() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The partition from which this record is received
- partition() - Method in exception org.apache.kafka.clients.consumer.NoOffsetForPartitionException
-
- partition(String, Object, byte[], Object, byte[], Cluster) - Method in interface org.apache.kafka.clients.producer.Partitioner
-
Compute the partition for the given record.
- partition() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- partition() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The partition the record was sent to
- partition(TopicPartition) - Method in class org.apache.kafka.common.Cluster
-
Get the metadata for the specified partition
- partition() - Method in class org.apache.kafka.common.PartitionInfo
-
The partition id
- partition() - Method in class org.apache.kafka.common.TopicPartition
-
- partition() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the partition id of the current input record; could be -1 if it is not
available (for example, if this method is invoked from the punctuate call)
- partition(K, V, int) - Method in interface org.apache.kafka.streams.processor.StreamPartitioner
-
Determine the partition number for a record with the given key and value and the current number of partitions.
- partition - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the partition.
- PARTITION_ASSIGNMENT_STRATEGY_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
partition.assignment.strategy
- PARTITION_GROUPER_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
partition.grouper
- partitionCountForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the number of partitions for the given topic
- Partitioner - Interface in org.apache.kafka.clients.producer
-
Partitioner Interface
- PARTITIONER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
partitioner.class
- PartitionGrouper - Interface in org.apache.kafka.streams.processor
-
A partition grouper that generates partition groups given the list of topic-partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in class org.apache.kafka.streams.processor.DefaultPartitionGrouper
-
Generate tasks with the assigned topic partitions.
- partitionGroups(Map<Integer, Set<String>>, Cluster) - Method in interface org.apache.kafka.streams.processor.PartitionGrouper
-
Returns a map of task ids to groups of partitions.
- PartitionInfo - Class in org.apache.kafka.common
-
Information about a topic-partition.
- PartitionInfo(String, int, Node, Node[], Node[]) - Constructor for class org.apache.kafka.common.PartitionInfo
-
- partitions() - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get the partitions which have records contained in this record set.
- partitions() - Method in exception org.apache.kafka.clients.consumer.InvalidOffsetException
-
- partitions() - Method in exception org.apache.kafka.clients.consumer.NoOffsetForPartitionException
-
- partitions() - Method in exception org.apache.kafka.clients.consumer.OffsetOutOfRangeException
-
- partitionsFor(String) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- partitionsFor(String) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get metadata about the partitions for a given topic.
- partitionsFor(String) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- partitionsFor(String) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Get the partition metadata for the give topic.
- partitionsFor(String) - Method in class org.apache.kafka.clients.producer.MockProducer
-
- partitionsFor(String) - Method in interface org.apache.kafka.clients.producer.Producer
-
Get a list of partitions for the given topic for custom partition assignment.
- partitionsForNode(int) - Method in class org.apache.kafka.common.Cluster
-
Get the list of partitions whose leader is this node
- partitionsForTopic(String) - Method in class org.apache.kafka.common.Cluster
-
Get the list of partitions for this topic
- pause(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- pause(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Suspend fetching from the requested partitions.
- pause(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- pause(TopicPartition...) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Pause consumption of messages from the specified TopicPartitions.
- paused() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- paused() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
- paused() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- peekNextKey() - Method in interface org.apache.kafka.streams.state.KeyValueIterator
-
Peek at the next key without advancing the iterator
- persistent() - Method in interface org.apache.kafka.streams.processor.StateStore
-
Return if the storage is persistent or not.
- persistent() - Method in interface org.apache.kafka.streams.state.Stores.KeyValueFactory
-
Keep all key-value entries off-heap in a local database, although for durability all entries are recorded in a Kafka
topic that can be read to restore the entries if they are lost.
- PolicyViolationException - Exception in org.apache.kafka.common.errors
-
Exception thrown if a create topics request does not satisfy the configured policy for a topic.
- PolicyViolationException(String) - Constructor for exception org.apache.kafka.common.errors.PolicyViolationException
-
- PolicyViolationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.PolicyViolationException
-
- poll(long) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- poll(long) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Fetch data for the topics or partitions specified using one of the subscribe/assign APIs.
- poll(long) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- poll() - Method in class org.apache.kafka.connect.source.SourceTask
-
Poll this SourceTask for new records.
- POLL_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
poll.ms
- port() - Method in class org.apache.kafka.common.Node
-
The port for this node
- port() - Method in class org.apache.kafka.streams.state.HostInfo
-
- port() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- position(TopicPartition) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- position(TopicPartition) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the offset of the next record that will be fetched (if a record with that offset exists).
- position(TopicPartition) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- preCommit(Map<TopicPartition, OffsetAndMetadata>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Pre-commit hook invoked prior to an offset commit.
- Predicate<K,V> - Interface in org.apache.kafka.streams.kstream
-
The
Predicate
interface represents a predicate (boolean-valued function) of a
KeyValue
pair.
- PRINCIPAL_BUILDER_CLASS_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- PRINCIPAL_BUILDER_CLASS_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- print() - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Print the records of this stream to System.out
.
- print() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the update records of this KTable
to System.out
.
- print(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the update records of this KTable
to System.out
.
- print(Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the update records of this KTable
to System.out
.
- print(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Print the update records of this KTable
to System.out
.
- process(ProcessorSupplier<? super K, ? super V>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
- process(K, V) - Method in interface org.apache.kafka.streams.processor.Processor
-
Process the record with the given key and value.
- Processor<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor of key-value pair records.
- ProcessorContext - Interface in org.apache.kafka.streams.processor
-
Processor context interface.
- ProcessorStateException - Exception in org.apache.kafka.streams.errors
-
Indicates a processor state operation (e.g.
- ProcessorStateException(String) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorStateException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.ProcessorStateException
-
- ProcessorSupplier<K,V> - Interface in org.apache.kafka.streams.processor
-
A processor supplier that can create one or more
Processor
instances.
- Producer<K,V> - Interface in org.apache.kafka.clients.producer
-
- PRODUCER_PREFIX - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- ProducerConfig - Class in org.apache.kafka.clients.producer
-
Configuration for the Kafka Producer.
- ProducerInterceptor<K,V> - Interface in org.apache.kafka.clients.producer
-
A plugin interface that allows you to intercept (and possibly mutate) the records received by the producer before
they are published to the Kafka cluster.
- producerPrefix(String) - Static method in class org.apache.kafka.streams.StreamsConfig
-
- ProducerRecord<K,V> - Class in org.apache.kafka.clients.producer
-
A key/value pair to be sent to Kafka.
- ProducerRecord(String, Integer, Long, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Creates a record with a specified timestamp to be sent to a specified topic and partition
- ProducerRecord(String, Integer, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Creates a record to be sent to a specified topic and partition
- ProducerRecord(String, K, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Create a record to be sent to Kafka
- ProducerRecord(String, V) - Constructor for class org.apache.kafka.clients.producer.ProducerRecord
-
Create a record with no key
- project(Schema, Object, Schema) - Static method in class org.apache.kafka.connect.data.SchemaProjector
-
This method project a value between compatible schemas and throw exceptions when non compatible schemas are provided
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
- punctuate(long) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
- punctuate(long) - Method in class org.apache.kafka.streams.processor.AbstractProcessor
-
- punctuate(long) - Method in interface org.apache.kafka.streams.processor.Processor
-
- put(String, Object) - Method in class org.apache.kafka.connect.data.Struct
-
Set the value of a field.
- put(Field, Object) - Method in class org.apache.kafka.connect.data.Struct
-
Set the value of a field.
- put(Collection<SinkRecord>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Put the records in the sink.
- put(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key
- put(Windowed<K>, AGG) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Write the aggregated value for the provided key to the store
- put(K, V) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the current wall-clock time as the timestamp
into the corresponding window
- put(K, V, long) - Method in interface org.apache.kafka.streams.state.WindowStore
-
Put a key-value pair with the given timestamp into the corresponding window
- putAll(List<KeyValue<K, V>>) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update all the given key/value pairs
- putIfAbsent(K, V) - Method in interface org.apache.kafka.streams.state.KeyValueStore
-
Update the value associated with this key, unless a value
is already associated with the key
- rack() - Method in class org.apache.kafka.common.Node
-
The rack for this node
- raiseError(Exception) - Method in interface org.apache.kafka.connect.connector.ConnectorContext
-
Raise an unrecoverable exception to the Connect framework.
- range(K, K) - Method in interface org.apache.kafka.streams.state.ReadOnlyKeyValueStore
-
Get an iterator over a given range of keys.
- RangeAssignor - Class in org.apache.kafka.clients.consumer
-
The range assignor works on a per-topic basis.
- RangeAssignor() - Constructor for class org.apache.kafka.clients.consumer.RangeAssignor
-
- rawKey(K) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given key.
- rawValue(V) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Serialize the given value.
- readFrom(DataInputStream) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- readFrom(ByteBuffer) - Static method in class org.apache.kafka.streams.processor.TaskId
-
- ReadOnlyKeyValueStore<K,V> - Interface in org.apache.kafka.streams.state
-
A key value store that only supports read operations.
- ReadOnlySessionStore<K,AGG> - Interface in org.apache.kafka.streams.state
-
A session store that only supports read operations.
- ReadOnlyWindowStore<K,V> - Interface in org.apache.kafka.streams.state
-
A window store that only supports read operations
Implementations should be thread-safe as concurrent reads and writes
are expected.
- rebalance(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
Simulate a rebalance event.
- RebalanceInProgressException - Exception in org.apache.kafka.common.errors
-
- RebalanceInProgressException() - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(String) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RebalanceInProgressException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RebalanceInProgressException
-
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
receive.buffer.bytes
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
receive.buffer.bytes
- RECEIVE_BUFFER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- recommendedValues() - Method in class org.apache.kafka.common.config.ConfigValue
-
- recommendedValues(List<Object>) - Method in class org.apache.kafka.common.config.ConfigValue
-
- recommender - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- reconfigure(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Reconfigure this Connector.
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
reconnect.backoff.ms
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
reconnect.backoff.ms
- RECONNECT_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- RecordBatchTooLargeException - Exception in org.apache.kafka.common.errors
-
This record batch is larger than the maximum allowable size
- RecordBatchTooLargeException() - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(String) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- RecordBatchTooLargeException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordBatchTooLargeException
-
- recordLatency(Sensor, long, long) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Record the given latency value of the sensor.
- RecordMetadata - Class in org.apache.kafka.clients.producer
-
The metadata for a record that has been acknowledged by the server
- RecordMetadata(TopicPartition, long, long) - Constructor for class org.apache.kafka.clients.producer.RecordMetadata
-
Deprecated.
- RecordMetadata(TopicPartition, long, long, long, long, int, int) - Constructor for class org.apache.kafka.clients.producer.RecordMetadata
-
- records(TopicPartition) - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get just the records for the given partition
- records(String) - Method in class org.apache.kafka.clients.consumer.ConsumerRecords
-
Get just the records for the given topic
- recordThroughput(Sensor, long) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Record the throughput value of a sensor.
- RecordTooLargeException - Exception in org.apache.kafka.common.errors
-
This record is larger than the maximum allowable size
- RecordTooLargeException() - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- RecordTooLargeException(String, Map<TopicPartition, Long>) - Constructor for exception org.apache.kafka.common.errors.RecordTooLargeException
-
- recordTooLargePartitions() - Method in exception org.apache.kafka.common.errors.RecordTooLargeException
-
- reduce(Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the values of records in this stream by the grouped key.
- reduce(Reducer<V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the value of records in this stream by the grouped key.
- reduce(Reducer<V>, Windows<W>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the number of records in this stream by the grouped key and the defined windows.
- reduce(Reducer<V>, Windows<W>, StateStoreSupplier<WindowStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
Combine the values of records in this stream by the grouped key and the defined windows.
- reduce(Reducer<V>, SessionWindows, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
- reduce(Reducer<V>, SessionWindows, StateStoreSupplier<SessionStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedStream
-
- reduce(Reducer<V>, Reducer<V>, String) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
.
- reduce(Reducer<V>, Reducer<V>, StateStoreSupplier<KeyValueStore>) - Method in interface org.apache.kafka.streams.kstream.KGroupedTable
-
Combine the value of records of the original
KTable
that got
mapped
to the same key into a new instance of
KTable
.
- Reducer<V> - Interface in org.apache.kafka.streams.kstream
-
The Reducer
interface for combining two values of the same type into a new value.
- register(StateStore, boolean, StateRestoreCallback) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Registers and possibly restores the specified storage engine.
- remove(Windowed<K>) - Method in interface org.apache.kafka.streams.state.SessionStore
-
Remove the session aggregated with provided
Windowed
key from the store
- removeSensor(Sensor) - Method in interface org.apache.kafka.streams.StreamsMetrics
-
Remove a sensor.
- repartitionSourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- ReplicaNotAvailableException - Exception in org.apache.kafka.common.errors
-
- ReplicaNotAvailableException(String) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- ReplicaNotAvailableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- ReplicaNotAvailableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.ReplicaNotAvailableException
-
- replicas() - Method in class org.apache.kafka.common.PartitionInfo
-
The complete set of replicas for this partition regardless of whether they are alive or up-to-date
- REPLICATION_FACTOR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
replication.factor
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
request.timeout.ms
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
request.timeout.ms
- REQUEST_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- requestCommit() - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Request an offset commit.
- requestTaskReconfiguration() - Method in interface org.apache.kafka.connect.connector.ConnectorContext
-
Requests that the runtime reconfigure the Tasks for this source.
- required() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set this schema as required.
- restore(byte[], byte[]) - Method in interface org.apache.kafka.streams.processor.StateRestoreCallback
-
- resume(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- resume(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
- resume(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- resume(TopicPartition...) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Resume consumption of messages from previously paused TopicPartitions.
- RetriableCommitFailedException - Exception in org.apache.kafka.clients.consumer
-
- RetriableCommitFailedException(Throwable) - Constructor for exception org.apache.kafka.clients.consumer.RetriableCommitFailedException
-
- RetriableCommitFailedException(String) - Constructor for exception org.apache.kafka.clients.consumer.RetriableCommitFailedException
-
- RetriableCommitFailedException(String, Throwable) - Constructor for exception org.apache.kafka.clients.consumer.RetriableCommitFailedException
-
- RetriableException - Exception in org.apache.kafka.common.errors
-
A retryable exception is a transient exception that if retried may succeed.
- RetriableException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException(String) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException(Throwable) - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException() - Constructor for exception org.apache.kafka.common.errors.RetriableException
-
- RetriableException - Exception in org.apache.kafka.connect.errors
-
An exception that indicates the operation can be reattempted.
- RetriableException(String) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RetriableException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RetriableException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.RetriableException
-
- RETRIES_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
retries
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
retry.backoff.ms
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
retry.backoff.ms
- RETRY_BACKOFF_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- ROCKSDB_CONFIG_SETTER_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
rocksdb.config.setter
- RocksDBConfigSetter - Interface in org.apache.kafka.streams.state
-
An interface to that allows developers to customize the RocksDB settings
for a given Store.
- RoundRobinAssignor - Class in org.apache.kafka.clients.consumer
-
The round robin assignor lays out all the available partitions and all the available consumers.
- RoundRobinAssignor() - Constructor for class org.apache.kafka.clients.consumer.RoundRobinAssignor
-
- SASL_ENABLED_MECHANISMS - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_ENABLED_MECHANISMS_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_JAAS_CONFIG - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_JAAS_CONFIG_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_KINIT_CMD - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_KINIT_CMD_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_MIN_TIME_BEFORE_RELOGIN - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_MIN_TIME_BEFORE_RELOGIN_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_PRINCIPAL_TO_LOCAL_RULES_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_SERVICE_NAME - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_SERVICE_NAME_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_JITTER - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_JITTER_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_WINDOW_FACTOR - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_KERBEROS_TICKET_RENEW_WINDOW_FACTOR_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SASL_MECHANISM - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
SASL mechanism configuration - standard mechanism names are listed
here.
- SASL_MECHANISM_DOC - Static variable in class org.apache.kafka.common.config.SaslConfigs
-
- SaslConfigs - Class in org.apache.kafka.common.config
-
- SaslConfigs() - Constructor for class org.apache.kafka.common.config.SaslConfigs
-
- SCALE_FIELD - Static variable in class org.apache.kafka.connect.data.Decimal
-
- schedule(long) - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Schedules a periodic operation for processors.
- scheduleNopPollTask() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- schedulePollTask(Runnable) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
Schedule a task to be executed during a poll().
- schema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Date
-
- schema(int) - Static method in class org.apache.kafka.connect.data.Decimal
-
- schema() - Method in class org.apache.kafka.connect.data.Field
-
Get the schema of this field
- Schema - Interface in org.apache.kafka.connect.data
-
Definition of an abstract data type.
- schema() - Method in interface org.apache.kafka.connect.data.Schema
-
Return a concrete instance of the
Schema
- schema() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- schema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Return a concrete instance of the
Schema
specified by this builder
- schema() - Method in class org.apache.kafka.connect.data.Struct
-
Get the schema for this Struct.
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Time
-
- SCHEMA - Static variable in class org.apache.kafka.connect.data.Timestamp
-
- Schema.Type - Enum in org.apache.kafka.connect.data
-
The type of a schema.
- SchemaAndValue - Class in org.apache.kafka.connect.data
-
- SchemaAndValue(Schema, Object) - Constructor for class org.apache.kafka.connect.data.SchemaAndValue
-
- SchemaBuilder - Class in org.apache.kafka.connect.data
-
SchemaBuilder provides a fluent API for constructing
Schema
objects.
- SchemaBuilderException - Exception in org.apache.kafka.connect.errors
-
- SchemaBuilderException(String) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaBuilderException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaBuilderException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaBuilderException
-
- SchemaProjector - Class in org.apache.kafka.connect.data
-
SchemaProjector is utility to project a value between compatible schemas and throw exceptions
when non compatible schemas are provided.
- SchemaProjector() - Constructor for class org.apache.kafka.connect.data.SchemaProjector
-
- SchemaProjectorException - Exception in org.apache.kafka.connect.errors
-
- SchemaProjectorException(String) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- SchemaProjectorException(String, Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- SchemaProjectorException(Throwable) - Constructor for exception org.apache.kafka.connect.errors.SchemaProjectorException
-
- schemaType(Class<?>) - Static method in class org.apache.kafka.connect.data.ConnectSchema
-
- SECURITY_PROTOCOL_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- seek(TopicPartition, long) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seek(TopicPartition, long) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Overrides the fetch offsets that the consumer will use on the next
poll(timeout)
.
- seek(TopicPartition, long) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- seekToBeginning(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seekToBeginning(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Seek to the first offset for each of the given partitions.
- seekToBeginning(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- seekToEnd(Collection<TopicPartition>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- seekToEnd(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Seek to the last offset for each of the given partitions.
- seekToEnd(Collection<TopicPartition>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- segments - Variable in class org.apache.kafka.streams.kstream.Windows
-
- segments(int) - Method in class org.apache.kafka.streams.kstream.Windows
-
Set the number of segments to be used for rolling the window store.
- selectKey(KeyValueMapper<? super K, ? super V, ? extends KR>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Set a new key (with possibly new type) for each input record.
- send(ProducerRecord<K, V>) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Asynchronously send a record to a topic.
- send(ProducerRecord<K, V>, Callback) - Method in class org.apache.kafka.clients.producer.KafkaProducer
-
Asynchronously send a record to a topic and invoke the provided callback when the send has been acknowledged.
- send(ProducerRecord<K, V>) - Method in class org.apache.kafka.clients.producer.MockProducer
-
Adds the record to the list of sent records.
- send(ProducerRecord<K, V>, Callback) - Method in class org.apache.kafka.clients.producer.MockProducer
-
Adds the record to the list of sent records.
- send(ProducerRecord<K, V>) - Method in interface org.apache.kafka.clients.producer.Producer
-
Send the given record asynchronously and return a future which will eventually contain the response information.
- send(ProducerRecord<K, V>, Callback) - Method in interface org.apache.kafka.clients.producer.Producer
-
Send a record and invoke the given callback when the record has been acknowledged by the server
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
send.buffer.bytes
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
send.buffer.bytes
- SEND_BUFFER_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
- Serde<T> - Interface in org.apache.kafka.common.serialization
-
The interface for wrapping a serializer and deserializer for the given data type.
- serdeFrom(Class<T>) - Static method in class org.apache.kafka.common.serialization.Serdes
-
- serdeFrom(Serializer<T>, Deserializer<T>) - Static method in class org.apache.kafka.common.serialization.Serdes
-
Construct a serde object from separate serializer and deserializer
- Serdes - Class in org.apache.kafka.common.serialization
-
Factory for creating serializers / deserializers.
- Serdes() - Constructor for class org.apache.kafka.common.serialization.Serdes
-
- Serdes.ByteArraySerde - Class in org.apache.kafka.common.serialization
-
- Serdes.ByteArraySerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.ByteArraySerde
-
- Serdes.ByteBufferSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.ByteBufferSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.ByteBufferSerde
-
- Serdes.BytesSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.BytesSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.BytesSerde
-
- Serdes.DoubleSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.DoubleSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.DoubleSerde
-
- Serdes.IntegerSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.IntegerSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.IntegerSerde
-
- Serdes.LongSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.LongSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.LongSerde
-
- Serdes.StringSerde - Class in org.apache.kafka.common.serialization
-
- Serdes.StringSerde() - Constructor for class org.apache.kafka.common.serialization.Serdes.StringSerde
-
- Serdes.WrapperSerde<T> - Class in org.apache.kafka.common.serialization
-
- Serdes.WrapperSerde(Serializer<T>, Deserializer<T>) - Constructor for class org.apache.kafka.common.serialization.Serdes.WrapperSerde
-
- SerializationException - Exception in org.apache.kafka.common.errors
-
Any exception during serialization in the producer
- SerializationException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException(String) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException(Throwable) - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- SerializationException() - Constructor for exception org.apache.kafka.common.errors.SerializationException
-
- serialize(String, byte[]) - Method in class org.apache.kafka.common.serialization.ByteArraySerializer
-
- serialize(String, ByteBuffer) - Method in class org.apache.kafka.common.serialization.ByteBufferSerializer
-
- serialize(String, Bytes) - Method in class org.apache.kafka.common.serialization.BytesSerializer
-
- serialize(String, Double) - Method in class org.apache.kafka.common.serialization.DoubleSerializer
-
- serialize(String, Integer) - Method in class org.apache.kafka.common.serialization.IntegerSerializer
-
- serialize(String, Long) - Method in class org.apache.kafka.common.serialization.LongSerializer
-
- serialize(String, T) - Method in interface org.apache.kafka.common.serialization.Serializer
-
- serialize(String, String) - Method in class org.apache.kafka.common.serialization.StringSerializer
-
- serializedKeySize() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The size of the serialized, uncompressed key in bytes.
- serializedKeySize() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The size of the serialized, uncompressed key in bytes.
- serializedValueSize() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The size of the serialized, uncompressed value in bytes.
- serializedValueSize() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The size of the serialized, uncompressed value in bytes.
- serializer() - Method in interface org.apache.kafka.common.serialization.Serde
-
- serializer() - Method in class org.apache.kafka.common.serialization.Serdes.WrapperSerde
-
- Serializer<T> - Interface in org.apache.kafka.common.serialization
-
- SESSION_TIMEOUT_MS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
session.timeout.ms
- sessionStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- SessionStore<K,AGG> - Interface in org.apache.kafka.streams.state
-
Interface for storing the aggregated values of sessions
- sessionWindowed(long) - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
- SessionWindows - Class in org.apache.kafka.streams.kstream
-
A session based window specification used for aggregating events into sessions.
- setApplicationId(String) - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Set the applicationId to be used for auto-generated internal topics.
- setConfig(String, Options, Map<String, Object>) - Method in interface org.apache.kafka.streams.state.RocksDBConfigSetter
-
Set the rocks db options for the provided storeName.
- setException(KafkaException) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- setStateListener(KafkaStreams.StateListener) - Method in class org.apache.kafka.streams.KafkaStreams
-
- setUncaughtExceptionHandler(Thread.UncaughtExceptionHandler) - Method in class org.apache.kafka.streams.KafkaStreams
-
Set the handler invoked when a
internal thread
abruptly
terminates due to an uncaught exception.
- SinkConnector - Class in org.apache.kafka.connect.sink
-
SinkConnectors implement the Connector interface to send Kafka data to another system.
- SinkConnector() - Constructor for class org.apache.kafka.connect.sink.SinkConnector
-
- SinkRecord - Class in org.apache.kafka.connect.sink
-
SinkRecord is a
ConnectRecord
that has been read from Kafka and includes the kafkaOffset of
the record in the Kafka topic-partition in addition to the standard fields.
- SinkRecord(String, int, Schema, Object, Schema, Object, long) - Constructor for class org.apache.kafka.connect.sink.SinkRecord
-
- SinkRecord(String, int, Schema, Object, Schema, Object, long, Long, TimestampType) - Constructor for class org.apache.kafka.connect.sink.SinkRecord
-
- SinkTask - Class in org.apache.kafka.connect.sink
-
SinkTask is a Task that takes records loaded from Kafka and sends them to another system.
- SinkTask() - Constructor for class org.apache.kafka.connect.sink.SinkTask
-
- SinkTaskContext - Interface in org.apache.kafka.connect.sink
-
Context passed to SinkTasks, allowing them to access utilities in the Kafka Connect runtime.
- sinkTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- size() - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
- size() - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
- size() - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return the size of the specified windows in milliseconds.
- size() - Method in class org.apache.kafka.streams.kstream.Windows
-
Return the size of the specified windows in milliseconds.
- sizeMs - Variable in class org.apache.kafka.streams.kstream.TimeWindows
-
The size of the windows in milliseconds.
- SourceConnector - Class in org.apache.kafka.connect.source
-
SourceConnectors implement the connector interface to pull data from another system and send
it to Kafka.
- SourceConnector() - Constructor for class org.apache.kafka.connect.source.SourceConnector
-
- sourceOffset() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- sourcePartition() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord - Class in org.apache.kafka.connect.source
-
SourceRecords are generated by SourceTasks and passed to Kafka Connect for storage in
Kafka.
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Integer, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Schema, Object, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Integer, Schema, Object, Schema, Object) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceRecord(Map<String, ?>, Map<String, ?>, String, Integer, Schema, Object, Schema, Object, Long) - Constructor for class org.apache.kafka.connect.source.SourceRecord
-
- SourceTask - Class in org.apache.kafka.connect.source
-
SourceTask is a Task that pulls records from another system for storage in Kafka.
- SourceTask() - Constructor for class org.apache.kafka.connect.source.SourceTask
-
- SourceTaskContext - Interface in org.apache.kafka.connect.source
-
SourceTaskContext is provided to SourceTasks to allow them to interact with the underlying
runtime.
- sourceTopicPattern() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- sourceTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- SSL_CIPHER_SUITES_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CIPHER_SUITES_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CLIENT_AUTH_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_CLIENT_AUTH_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENABLED_PROTOCOLS_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENABLED_PROTOCOLS_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENDPOINT_IDENTIFICATION_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_ENDPOINT_IDENTIFICATION_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEY_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEY_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYMANAGER_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYMANAGER_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_LOCATION_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_LOCATION_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_TYPE_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_KEYSTORE_TYPE_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROTOCOL_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROTOCOL_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROVIDER_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_PROVIDER_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_SECURE_RANDOM_IMPLEMENTATION_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_SECURE_RANDOM_IMPLEMENTATION_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTMANAGER_ALGORITHM_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTMANAGER_ALGORITHM_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_LOCATION_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_LOCATION_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_PASSWORD_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_PASSWORD_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_TYPE_CONFIG - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SSL_TRUSTSTORE_TYPE_DOC - Static variable in class org.apache.kafka.common.config.SslConfigs
-
- SslConfigs - Class in org.apache.kafka.common.config
-
- SslConfigs() - Constructor for class org.apache.kafka.common.config.SslConfigs
-
- start(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Start this Connector.
- start(Map<String, String>) - Method in interface org.apache.kafka.connect.connector.Task
-
Start the Task
- start(Map<String, String>) - Method in class org.apache.kafka.connect.sink.SinkTask
-
Start the Task.
- start(Map<String, String>) - Method in class org.apache.kafka.connect.source.SourceTask
-
Start the Task.
- start() - Method in class org.apache.kafka.streams.KafkaStreams
-
Start the KafkaStreams
instance by starting all its threads.
- start() - Method in class org.apache.kafka.streams.kstream.Window
-
Return the start timestamp of this window.
- startMs - Variable in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
The start timestamp of the window.
- startMs - Variable in class org.apache.kafka.streams.kstream.Window
-
- startOn(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
Return a new unlimited window for the specified start timestamp.
- state() - Method in class org.apache.kafka.streams.KafkaStreams
-
- STATE_CLEANUP_DELAY_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.cleanup.delay
- STATE_DIR_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
state.dir
- stateChangelogTopics - Variable in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- stateDir() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the state directory for the partition.
- StateRestoreCallback - Interface in org.apache.kafka.streams.processor
-
Restoration logic for log-backed state stores upon restart,
it takes one record at a time from the logs to apply to the restoring state.
- StateSerdes<K,V> - Class in org.apache.kafka.streams.state
-
Factory for creating serializers / deserializers for state stores in Kafka Streams.
- StateSerdes(String, Serde<K>, Serde<V>) - Constructor for class org.apache.kafka.streams.state.StateSerdes
-
Create a context for serialization using the specified serializers and deserializers which
must match the key and value types used as parameters for this object; the state changelog topic
is provided to bind this serde factory to, so that future calls for serialize / deserialize do not
need to provide the topic name any more.
- StateStore - Interface in org.apache.kafka.streams.processor
-
A storage engine for managing state maintained by a stream processor.
- stateStoreNames() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- stateStoreNameToSourceTopics() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
- StateStoreSupplier<T extends StateStore> - Interface in org.apache.kafka.streams.processor
-
A state store supplier which can create one or more
StateStore
instances.
- stop() - Method in class org.apache.kafka.connect.connector.Connector
-
Stop this connector.
- stop() - Method in interface org.apache.kafka.connect.connector.Task
-
Stop this task.
- stop() - Method in class org.apache.kafka.connect.sink.SinkTask
-
Perform any cleanup to stop this task.
- stop() - Method in class org.apache.kafka.connect.source.SourceTask
-
Signal this SourceTask to stop.
- store(String, QueryableStoreType<T>) - Method in class org.apache.kafka.streams.KafkaStreams
-
Get a facade wrapping the local
StateStore
instances with the provided
storeName
if the Store's
type is accepted by the provided
queryableStoreType
.
- Stores - Class in org.apache.kafka.streams.state
-
Factory for creating state stores in Kafka Streams.
- Stores() - Constructor for class org.apache.kafka.streams.state.Stores
-
- Stores.InMemoryKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create in-memory key-value stores.
- Stores.KeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
- Stores.PersistentKeyValueFactory<K,V> - Interface in org.apache.kafka.streams.state
-
The interface used to create off-heap key-value stores that use a local database.
- Stores.StoreFactory - Class in org.apache.kafka.streams.state
-
- Stores.StoreFactory() - Constructor for class org.apache.kafka.streams.state.Stores.StoreFactory
-
- Stores.ValueFactory<K> - Class in org.apache.kafka.streams.state
-
The factory for creating off-heap key-value stores.
- Stores.ValueFactory() - Constructor for class org.apache.kafka.streams.state.Stores.ValueFactory
-
- stream(String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TopologyBuilder.AutoOffsetReset, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TopologyBuilder.AutoOffsetReset, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, String...) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topics.
- stream(Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- stream(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, Pattern) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KStream
from the specified topic pattern.
- STREAM_THREAD_INSTANCE - Static variable in class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamPartitioner<K,V> - Interface in org.apache.kafka.streams.processor
-
Determine how records are distributed among the partitions in a Kafka topic.
- StreamsConfig - Class in org.apache.kafka.streams
-
- StreamsConfig(Map<?, ?>) - Constructor for class org.apache.kafka.streams.StreamsConfig
-
Create a new StreamsConfig
using the given properties.
- StreamsConfig.InternalConfig - Class in org.apache.kafka.streams
-
- StreamsConfig.InternalConfig() - Constructor for class org.apache.kafka.streams.StreamsConfig.InternalConfig
-
- StreamsException - Exception in org.apache.kafka.streams.errors
-
- StreamsException(String) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.StreamsException
-
- StreamsMetadata - Class in org.apache.kafka.streams.state
-
Represents the state of an instance (process) in a
KafkaStreams
application.
- StreamsMetadata(HostInfo, Set<String>, Set<TopicPartition>) - Constructor for class org.apache.kafka.streams.state.StreamsMetadata
-
- StreamsMetrics - Interface in org.apache.kafka.streams
-
The Kafka Streams metrics interface for adding metric sensors and collecting metric values.
- String() - Static method in class org.apache.kafka.common.serialization.Serdes
-
- string() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- STRING_SCHEMA - Static variable in interface org.apache.kafka.connect.data.Schema
-
- StringConverter - Class in org.apache.kafka.connect.storage
-
Converter
implementation that only supports serializing to strings.
- StringConverter() - Constructor for class org.apache.kafka.connect.storage.StringConverter
-
- StringDeserializer - Class in org.apache.kafka.common.serialization
-
String encoding defaults to UTF8 and can be customized by setting the property key.deserializer.encoding,
value.deserializer.encoding or deserializer.encoding.
- StringDeserializer() - Constructor for class org.apache.kafka.common.serialization.StringDeserializer
-
- StringSerializer - Class in org.apache.kafka.common.serialization
-
String encoding defaults to UTF8 and can be customized by setting the property key.serializer.encoding,
value.serializer.encoding or serializer.encoding.
- StringSerializer() - Constructor for class org.apache.kafka.common.serialization.StringSerializer
-
- struct() - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
- Struct - Class in org.apache.kafka.connect.data
-
A structured record containing a set of named fields with values, each field using an independent
Schema
.
- Struct(Schema) - Constructor for class org.apache.kafka.connect.data.Struct
-
Create a new Struct for this
Schema
- subscribe(Collection<String>) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Pattern, ConsumerRebalanceListener) - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to the given list of topics to get dynamically
assigned partitions.
- subscribe(Collection<String>) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to the given list of topics to get dynamically assigned partitions.
- subscribe(Pattern, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Subscribe to all topics matching specified pattern to get dynamically assigned partitions.
- subscribe(Collection<String>) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscribe(Pattern, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscribe(Collection<String>, ConsumerRebalanceListener) - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- subscription() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- subscription() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Get the current subscription.
- subscription() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- table(String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- table(TopologyBuilder.AutoOffsetReset, Serde<K>, Serde<V>, String, String) - Method in class org.apache.kafka.streams.kstream.KStreamBuilder
-
Create a
KTable
for the specified topic.
- tags() - Method in class org.apache.kafka.common.MetricName
-
- Task - Interface in org.apache.kafka.connect.connector
-
Tasks contain the code that actually copies data to/from another system.
- TaskAssignmentException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying to assign
stream tasks
to
threads
.
- TaskAssignmentException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- TaskAssignmentException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskAssignmentException
-
- taskClass() - Method in class org.apache.kafka.connect.connector.Connector
-
Returns the Task implementation for this Connector.
- taskConfigs(int) - Method in class org.apache.kafka.connect.connector.Connector
-
Returns a set of configurations for Tasks based on the current configuration,
producing at most count configurations.
- taskId() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the task id
- TaskId - Class in org.apache.kafka.streams.processor
-
The task ID representation composed as topic group ID plus the assigned partition ID.
- TaskId(int, int) - Constructor for class org.apache.kafka.streams.processor.TaskId
-
- TaskIdFormatException - Exception in org.apache.kafka.streams.errors
-
Indicates a run time error incurred while trying parse the
task id
from the read string.
- TaskIdFormatException(String) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- TaskIdFormatException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- TaskIdFormatException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TaskIdFormatException
-
- test(K, V) - Method in interface org.apache.kafka.streams.kstream.Predicate
-
Test if the record with the given key and value satisfies the predicate.
- through(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new KStream
from the topic using default serializers and
deserializers and producer's DefaultPartitioner
.
- through(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new
KStream
from the topic using default serializers and
deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic, and creates a new KStream
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic and creates a new
KStream
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- through(String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic using default
serializers and deserializers and producer's DefaultPartitioner
.
- through(StreamPartitioner<? super K, ? super V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using default
serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of
records to partitions.
- through(Serde<K>, Serde<V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new KTable
from the topic.
- through(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic and creates a new
KTable
from the topic using a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- Time - Class in org.apache.kafka.connect.data
-
A time representing a specific point in a day, not tied to any specific date.
- Time() - Constructor for class org.apache.kafka.connect.data.Time
-
- timeout(long) - Method in interface org.apache.kafka.connect.sink.SinkTaskContext
-
Set the timeout in milliseconds.
- TIMEOUT_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- TimeoutException - Exception in org.apache.kafka.common.errors
-
Indicates that a request timed out.
- TimeoutException() - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(String) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- TimeoutException(Throwable) - Constructor for exception org.apache.kafka.common.errors.TimeoutException
-
- timestamp() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The timestamp of this record
- timestamp() - Method in class org.apache.kafka.clients.consumer.OffsetAndTimestamp
-
- timestamp() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- timestamp() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The timestamp of the record in the topic/partition.
- timestamp() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- Timestamp - Class in org.apache.kafka.connect.data
-
A timestamp representing an absolute time, without timezone information.
- Timestamp() - Constructor for class org.apache.kafka.connect.data.Timestamp
-
- timestamp() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the current timestamp.
- TIMESTAMP_EXTRACTOR_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
timestamp.extractor
- TimestampExtractor - Interface in org.apache.kafka.streams.processor
-
An interface that allows the Kafka Streams framework to extract a timestamp from an instance of
ConsumerRecord
.
- timestampType() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The timestamp type of this record
- timestampType() - Method in class org.apache.kafka.connect.sink.SinkRecord
-
- TimeWindows - Class in org.apache.kafka.streams.kstream
-
The fixed-size time-based window specifications used for aggregations.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config and producer's
DefaultPartitioner
.
- to(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using default serializers specified in the config and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Materialize this stream to a topic using a customizable
StreamPartitioner
to determine the distribution
of records to partitions.
- to(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using default serializers and deserializers and producer's
DefaultPartitioner
.
- to(StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using default serializers and deserializers and a customizable
StreamPartitioner
to determine the distribution of records to partitions.
- to(Serde<K>, Serde<V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic.
- to(Serde<K>, Serde<V>, StreamPartitioner<? super K, ? super V>, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Materialize this changelog stream to a topic using a customizable
StreamPartitioner
to determine the
distribution of records to partitions.
- toConnectData(String, byte[]) - Method in interface org.apache.kafka.connect.storage.Converter
-
Convert a native object to a Kafka Connect data object.
- toConnectData(String, byte[]) - Method in class org.apache.kafka.connect.storage.StringConverter
-
- toEnrichedRst() - Method in class org.apache.kafka.common.config.ConfigDef
-
Configs with new metadata (group, orderInGroup, dependents) formatted with reStructuredText, suitable for embedding in Sphinx
documentation.
- toHtmlTable() - Method in class org.apache.kafka.common.config.ConfigDef
-
- toLogical(Schema, int) - Static method in class org.apache.kafka.connect.data.Date
-
- toLogical(Schema, byte[]) - Static method in class org.apache.kafka.connect.data.Decimal
-
- toLogical(Schema, int) - Static method in class org.apache.kafka.connect.data.Time
-
- toLogical(Schema, long) - Static method in class org.apache.kafka.connect.data.Timestamp
-
- topic() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The topic this record is received from
- topic() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- topic() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
The topic the record was appended to
- topic() - Method in class org.apache.kafka.common.PartitionInfo
-
The topic name
- topic() - Method in class org.apache.kafka.common.TopicPartition
-
- topic() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- topic() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the topic name of the current input record; could be null if it is not
available (for example, if this method is invoked from the punctuate call)
- topic() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the topic.
- TopicAuthorizationException - Exception in org.apache.kafka.common.errors
-
- TopicAuthorizationException(Set<String>) - Constructor for exception org.apache.kafka.common.errors.TopicAuthorizationException
-
- TopicAuthorizationException(String) - Constructor for exception org.apache.kafka.common.errors.TopicAuthorizationException
-
- TopicExistsException - Exception in org.apache.kafka.common.errors
-
- TopicExistsException(String) - Constructor for exception org.apache.kafka.common.errors.TopicExistsException
-
- TopicExistsException(String, Throwable) - Constructor for exception org.apache.kafka.common.errors.TopicExistsException
-
- topicGroupId - Variable in class org.apache.kafka.streams.processor.TaskId
-
The ID of the topic group.
- topicGroups() - Method in class org.apache.kafka.streams.processor.TopologyBuilder
-
Returns the map of topic groups keyed by the group id.
- TopicPartition - Class in org.apache.kafka.common
-
A topic name and partition number
- TopicPartition(String, int) - Constructor for class org.apache.kafka.common.TopicPartition
-
- topicPartitions() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- topics() - Method in class org.apache.kafka.common.Cluster
-
Get all topics.
- TOPICS_CONFIG - Static variable in class org.apache.kafka.connect.sink.SinkConnector
-
Configuration key for the list of input topics for this connector.
- TOPICS_CONFIG - Static variable in class org.apache.kafka.connect.sink.SinkTask
-
The configuration key that provides the list of topics that are inputs for this
SinkTask.
- TopologyBuilder - Class in org.apache.kafka.streams.processor
-
A component that is used to build a ProcessorTopology
.
- TopologyBuilder() - Constructor for class org.apache.kafka.streams.processor.TopologyBuilder
-
Create a new builder.
- TopologyBuilder.AutoOffsetReset - Enum in org.apache.kafka.streams.processor
-
Enum used to define auto offset reset policy when creating
KStream
or
KTable
- TopologyBuilder.TopicsInfo - Class in org.apache.kafka.streams.processor
-
- TopologyBuilderException - Exception in org.apache.kafka.streams.errors
-
Indicates a pre-run time error incurred while parsing the
builder
to construct the
processor topology
.
- TopologyBuilderException(String) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- TopologyBuilderException(String, Throwable) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- TopologyBuilderException(Throwable) - Constructor for exception org.apache.kafka.streams.errors.TopologyBuilderException
-
- toRst() - Method in class org.apache.kafka.common.config.ConfigDef
-
Get the configs formatted with reStructuredText, suitable for embedding in Sphinx
documentation.
- toStream() - Method in interface org.apache.kafka.streams.kstream.KTable
-
Convert this changelog stream to a
KStream
.
- toStream(KeyValueMapper<? super K, ? super V, ? extends KR>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
- toString() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
- toString() - Method in class org.apache.kafka.clients.consumer.OffsetAndMetadata
-
- toString() - Method in class org.apache.kafka.clients.consumer.OffsetAndTimestamp
-
- toString() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- toString() - Method in class org.apache.kafka.clients.producer.RecordMetadata
-
- toString() - Method in class org.apache.kafka.common.Cluster
-
- toString() - Method in class org.apache.kafka.common.ClusterResource
-
- toString() - Method in class org.apache.kafka.common.config.ConfigDef.Range
-
- toString() - Method in class org.apache.kafka.common.config.ConfigDef.ValidList
-
- toString() - Method in class org.apache.kafka.common.config.ConfigDef.ValidString
-
- toString() - Method in class org.apache.kafka.common.config.ConfigValue
-
- toString() - Method in class org.apache.kafka.common.MetricName
-
- toString() - Method in class org.apache.kafka.common.Node
-
- toString() - Method in class org.apache.kafka.common.PartitionInfo
-
- toString() - Method in class org.apache.kafka.common.TopicPartition
-
- toString() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- toString() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- toString() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- toString() - Method in class org.apache.kafka.connect.data.Struct
-
- toString() - Method in class org.apache.kafka.connect.sink.SinkRecord
-
- toString() - Method in class org.apache.kafka.connect.source.SourceRecord
-
- toString() - Method in class org.apache.kafka.streams.KafkaStreams
-
Produce a string representation containing useful information about this
KafkaStream
instance such as
thread IDs, task IDs, and a representation of the topology DAG including
StateStore
s (cf.
- toString(String) - Method in class org.apache.kafka.streams.KafkaStreams
-
Produce a string representation containing useful information about this
KafkaStream
instance such as
thread IDs, task IDs, and a representation of the topology DAG including
StateStore
s (cf.
- toString() - Method in class org.apache.kafka.streams.KeyValue
-
- toString() - Method in class org.apache.kafka.streams.kstream.Window
-
- toString() - Method in class org.apache.kafka.streams.kstream.Windowed
-
- toString() - Method in class org.apache.kafka.streams.processor.TaskId
-
- toString() - Method in class org.apache.kafka.streams.processor.TopologyBuilder.TopicsInfo
-
- toString() - Method in class org.apache.kafka.streams.state.HostInfo
-
- toString() - Method in class org.apache.kafka.streams.state.StreamsMetadata
-
- transform(TransformerSupplier<? super K, ? super V, KeyValue<K1, V1>>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform each record of the input stream into zero or more records in the output stream (both key and value type
can be altered arbitrarily).
- transform(K, V) - Method in interface org.apache.kafka.streams.kstream.Transformer
-
Transform the record with the given key and value.
- transform(V) - Method in interface org.apache.kafka.streams.kstream.ValueTransformer
-
Transform the given value to a new value.
- Transformation<R extends ConnectRecord<R>> - Interface in org.apache.kafka.connect.transforms
-
Single message transformation for Kafka Connect record types.
- Transformer<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
The Transformer
interface for stateful mapping of an input record to zero, one, or multiple new output
records (both key and value type can be altered arbitrarily).
- TransformerSupplier<K,V,R> - Interface in org.apache.kafka.streams.kstream
-
A
TransformerSupplier
interface which can create one or more
Transformer
instances.
- transformValues(ValueTransformerSupplier<? super V, ? extends VR>, String...) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Transform the value of each input record into a new value (with possible new type) of the output record.
- type - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- type() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- type() - Method in interface org.apache.kafka.connect.data.Schema
-
- type() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- type(Schema.Type) - Static method in class org.apache.kafka.connect.data.SchemaBuilder
-
Create a SchemaBuilder for the specified type.
- validate(Map<String, String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
Validate the current configuration values with the configuration definition.
- validate(Map<String, String>) - Method in class org.apache.kafka.connect.connector.Connector
-
Validate the connector configuration values against configuration definitions.
- validate() - Method in class org.apache.kafka.connect.data.Struct
-
Validates that this struct has filled in all the necessary data with valid values.
- validateAll(Map<String, String>) - Method in class org.apache.kafka.common.config.ConfigDef
-
- validateValue(Schema, Object) - Static method in class org.apache.kafka.connect.data.ConnectSchema
-
Validate that the value can be used with the schema, i.e.
- validateValue(Object) - Method in class org.apache.kafka.connect.data.ConnectSchema
-
Validate that the value can be used for this schema, i.e.
- validator - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- validValues(String, Map<String, Object>) - Method in interface org.apache.kafka.common.config.ConfigDef.Recommender
-
The valid values for the configuration given the current configuration values.
- value() - Method in class org.apache.kafka.clients.consumer.ConsumerRecord
-
The value
- value() - Method in class org.apache.kafka.clients.producer.ProducerRecord
-
- value() - Method in class org.apache.kafka.common.config.ConfigValue
-
- value(Object) - Method in class org.apache.kafka.common.config.ConfigValue
-
- value() - Method in interface org.apache.kafka.common.Metric
-
The value of the metric
- value() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- value() - Method in class org.apache.kafka.connect.data.SchemaAndValue
-
- value - Variable in class org.apache.kafka.streams.KeyValue
-
The value of the key-value pair.
- VALUE_DESERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
value.deserializer
- VALUE_DESERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.consumer.ConsumerConfig
-
- VALUE_SERDE_CLASS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
value.serde
- VALUE_SERIALIZER_CLASS_CONFIG - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
value.serializer
- VALUE_SERIALIZER_CLASS_DOC - Static variable in class org.apache.kafka.clients.producer.ProducerConfig
-
- valueDeserializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value deserializer.
- valueFrom(byte[]) - Method in class org.apache.kafka.streams.state.StateSerdes
-
Deserialize the value from raw bytes.
- ValueJoiner<V1,V2,VR> - Interface in org.apache.kafka.streams.kstream
-
The ValueJoiner
interface for joining two values into a new value of arbitrary type.
- ValueMapper<V,VR> - Interface in org.apache.kafka.streams.kstream
-
The ValueMapper
interface for mapping a value to a new value of arbitrary type.
- valueOf(String) - Static method in enum org.apache.kafka.clients.consumer.OffsetResetStrategy
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Importance
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Type
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.common.config.ConfigDef.Width
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.connect.data.Schema.Type
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.streams.KafkaStreams.State
-
Returns the enum constant of this type with the specified name.
- valueOf(String) - Static method in enum org.apache.kafka.streams.processor.TopologyBuilder.AutoOffsetReset
-
Returns the enum constant of this type with the specified name.
- values() - Static method in enum org.apache.kafka.clients.consumer.OffsetResetStrategy
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Method in class org.apache.kafka.common.config.AbstractConfig
-
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Importance
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Type
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.common.config.ConfigDef.Width
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.connect.data.Schema.Type
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.streams.KafkaStreams.State
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- values() - Static method in enum org.apache.kafka.streams.processor.TopologyBuilder.AutoOffsetReset
-
Returns an array containing the constants of this enum type, in
the order they are declared.
- valueSchema() - Method in class org.apache.kafka.connect.connector.ConnectRecord
-
- valueSchema() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- valueSchema() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the value schema for this map or array schema.
- valueSchema() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- valueSerde() - Method in interface org.apache.kafka.streams.processor.ProcessorContext
-
Returns the default value serde
- valueSerde() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value serde.
- valueSerde() - Method in class org.apache.kafka.streams.StreamsConfig
-
- valueSerializer() - Method in class org.apache.kafka.streams.state.StateSerdes
-
Return the value serializer.
- ValueTransformer<V,VR> - Interface in org.apache.kafka.streams.kstream
-
The ValueTransformer
interface for stateful mapping of a value to a new value (with possible new type).
- ValueTransformerSupplier<V,VR> - Interface in org.apache.kafka.streams.kstream
-
A
ValueTransformerSupplier
interface which can create one or more
ValueTransformer
instances.
- version() - Method in class org.apache.kafka.connect.connector.Connector
-
Get the version of this connector.
- version() - Method in interface org.apache.kafka.connect.connector.Task
-
Get the version of this task.
- version() - Method in class org.apache.kafka.connect.data.ConnectSchema
-
- version() - Method in interface org.apache.kafka.connect.data.Schema
-
Get the optional version of the schema.
- version() - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
- version(Integer) - Method in class org.apache.kafka.connect.data.SchemaBuilder
-
Set the version of this schema.
- visible(String, Map<String, Object>) - Method in interface org.apache.kafka.common.config.ConfigDef.Recommender
-
Set the visibility of the configuration given the current configuration values.
- visible() - Method in class org.apache.kafka.common.config.ConfigValue
-
- visible(boolean) - Method in class org.apache.kafka.common.config.ConfigValue
-
- wakeup() - Method in interface org.apache.kafka.clients.consumer.Consumer
-
- wakeup() - Method in class org.apache.kafka.clients.consumer.KafkaConsumer
-
Wakeup the consumer.
- wakeup() - Method in class org.apache.kafka.clients.consumer.MockConsumer
-
- WakeupException - Exception in org.apache.kafka.common.errors
-
Exception used to indicate preemption of a blocking operation by an external thread.
- WakeupException() - Constructor for exception org.apache.kafka.common.errors.WakeupException
-
- WallclockTimestampExtractor - Class in org.apache.kafka.streams.processor
-
- WallclockTimestampExtractor() - Constructor for class org.apache.kafka.streams.processor.WallclockTimestampExtractor
-
- width - Variable in class org.apache.kafka.common.config.ConfigDef.ConfigKey
-
- Window - Class in org.apache.kafka.streams.kstream
-
A single window instance, defined by its start and end timestamp.
- Window(long, long) - Constructor for class org.apache.kafka.streams.kstream.Window
-
Create a new window for the given start and end time.
- window() - Method in class org.apache.kafka.streams.kstream.Windowed
-
Return the window containing the values associated with this key.
- WINDOW_STORE_CHANGE_LOG_ADDITIONAL_RETENTION_MS_CONFIG - Static variable in class org.apache.kafka.streams.StreamsConfig
-
windowstore.changelog.additional.retention.ms
- Windowed<K> - Class in org.apache.kafka.streams.kstream
-
The result key type of a windowed stream aggregation.
- Windowed(K, Window) - Constructor for class org.apache.kafka.streams.kstream.Windowed
-
- windowed(long, long, int, boolean) - Method in interface org.apache.kafka.streams.state.Stores.PersistentKeyValueFactory
-
Set the persistent store as a windowed key-value store
- Windows<W extends Window> - Class in org.apache.kafka.streams.kstream
-
The window specification interface for fixed size windows that is used to define window boundaries and window
maintain duration.
- Windows() - Constructor for class org.apache.kafka.streams.kstream.Windows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.JoinWindows
-
Not supported by JoinWindows
.
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.TimeWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.UnlimitedWindows
-
- windowsFor(long) - Method in class org.apache.kafka.streams.kstream.Windows
-
Create all windows that contain the provided timestamp, indexed by non-negative window start timestamps.
- windowStore() - Static method in class org.apache.kafka.streams.state.QueryableStoreTypes
-
- WindowStore<K,V> - Interface in org.apache.kafka.streams.state
-
- WindowStoreIterator<V> - Interface in org.apache.kafka.streams.state
-
- with(long) - Static method in class org.apache.kafka.streams.kstream.SessionWindows
-
Create a new window specification with the specified inactivity gap in milliseconds.
- withBuiltinTypes(String, Class<K>, Class<V>) - Static method in class org.apache.kafka.streams.state.StateSerdes
-
Create a new instance of
StateSerdes
for the given state name and key-/value-type classes.
- withByteArrayKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the keys will be byte arrays.
- withByteArrayValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use byte arrays for values.
- withByteBufferKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withByteBufferValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withClientSaslSupport() - Method in class org.apache.kafka.common.config.ConfigDef
-
Add standard SASL client configuration options.
- withClientSslSupport() - Method in class org.apache.kafka.common.config.ConfigDef
-
Add standard SSL client configuration options.
- withDoubleKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withDoubleValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withIntegerKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withIntegerValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withKeys(Class<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withKeys(Serde<K>) - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
Begin to create a
KeyValueStore
by specifying the serializer and deserializer for the keys.
- withLongKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withLongValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withPartitions(Map<TopicPartition, PartitionInfo>) - Method in class org.apache.kafka.common.Cluster
-
Return a copy of this cluster combined with `partitions`.
- withStringKeys() - Method in class org.apache.kafka.streams.state.Stores.StoreFactory
-
- withStringValues() - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
- withValues(Class<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use values of the specified type.
- withValues(Serde<V>) - Method in class org.apache.kafka.streams.state.Stores.ValueFactory
-
Use the specified serializer and deserializer for the values.
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, String) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String, String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KStream
-
Write the records of this stream to a file at the given path.
- writeAsText(String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the update records of this KTable
to a file at the given path.
- writeAsText(String, String) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the update records of this KTable
to a file at the given path.
- writeAsText(String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the update records of this KTable
to a file at the given path.
- writeAsText(String, String, Serde<K>, Serde<V>) - Method in interface org.apache.kafka.streams.kstream.KTable
-
Write the update records of this KTable
to a file at the given path.
- writeTo(DataOutputStream) - Method in class org.apache.kafka.streams.processor.TaskId
-
- writeTo(ByteBuffer) - Method in class org.apache.kafka.streams.processor.TaskId
-