You are viewing documentation for an older version (3.9) of Kafka. For up-to-date documentation, see the latest version.
MirrorMaker Configs
Below is the configuration of the connectors that make up MirrorMaker 2.
MirrorMaker Common Configs
Below are the common configuration properties that apply to all three connectors.
source.cluster.alias
Alias of source cluster
Type: string Default: Valid Values: Importance: high ssl.key.password
The password of the private key in the key store file or the PEM key specified in 'ssl.keystore.key'.
Type: password Default: null Valid Values: Importance: high ssl.keystore.certificate.chain
Certificate chain in the format specified by 'ssl.keystore.type'. Default SSL engine factory supports only PEM format with a list of X.509 certificates
Type: password Default: null Valid Values: Importance: high ssl.keystore.key
Private key in the format specified by 'ssl.keystore.type'. Default SSL engine factory supports only PEM format with PKCS#8 keys. If the key is encrypted, key password must be specified using 'ssl.key.password'
Type: password Default: null Valid Values: Importance: high ssl.keystore.location
The location of the key store file. This is optional for client and can be used for two-way authentication for client.
Type: string Default: null Valid Values: Importance: high ssl.keystore.password
The store password for the key store file. This is optional for client and only needed if 'ssl.keystore.location' is configured. Key store password is not supported for PEM format.
Type: password Default: null Valid Values: Importance: high ssl.truststore.certificates
Trusted certificates in the format specified by 'ssl.truststore.type'. Default SSL engine factory supports only PEM format with X.509 certificates.
Type: password Default: null Valid Values: Importance: high ssl.truststore.location
The location of the trust store file.
Type: string Default: null Valid Values: Importance: high ssl.truststore.password
The password for the trust store file. If a password is not set, trust store file configured will still be used, but integrity checking is disabled. Trust store password is not supported for PEM format.
Type: password Default: null Valid Values: Importance: high target.cluster.alias
Alias of target cluster. Used in metrics reporting.
Type: string Default: target Valid Values: Importance: high sasl.client.callback.handler.class
The fully qualified name of a SASL client callback handler class that implements the AuthenticateCallbackHandler interface.
Type: class Default: null Valid Values: Importance: medium sasl.jaas.config
JAAS login context parameters for SASL connections in the format used by JAAS configuration files. JAAS configuration file format is described here. The format for the value is:
loginModuleClass controlFlag (optionName=optionValue)*;. For brokers, the config must be prefixed with listener prefix and SASL mechanism name in lower-case. For example, listener.name.sasl_ssl.scram-sha-256.sasl.jaas.config=com.example.ScramLoginModule required;Type: password Default: null Valid Values: Importance: medium sasl.kerberos.service.name
The Kerberos principal name that Kafka runs as. This can be defined either in Kafka's JAAS config or in Kafka's config.
Type: string Default: null Valid Values: Importance: medium sasl.login.callback.handler.class
The fully qualified name of a SASL login callback handler class that implements the AuthenticateCallbackHandler interface. For brokers, login callback handler config must be prefixed with listener prefix and SASL mechanism name in lower-case. For example, listener.name.sasl_ssl.scram-sha-256.sasl.login.callback.handler.class=com.example.CustomScramLoginCallbackHandler
Type: class Default: null Valid Values: Importance: medium sasl.login.class
The fully qualified name of a class that implements the Login interface. For brokers, login config must be prefixed with listener prefix and SASL mechanism name in lower-case. For example, listener.name.sasl_ssl.scram-sha-256.sasl.login.class=com.example.CustomScramLogin
Type: class Default: null Valid Values: Importance: medium sasl.mechanism
SASL mechanism used for client connections. This may be any mechanism for which a security provider is available. GSSAPI is the default mechanism.
Type: string Default: GSSAPI Valid Values: Importance: medium sasl.oauthbearer.jwks.endpoint.url
The OAuth/OIDC provider URL from which the provider's JWKS (JSON Web Key Set) can be retrieved. The URL can be HTTP(S)-based or file-based. If the URL is HTTP(S)-based, the JWKS data will be retrieved from the OAuth/OIDC provider via the configured URL on broker startup. All then-current keys will be cached on the broker for incoming requests. If an authentication request is received for a JWT that includes a "kid" header claim value that isn't yet in the cache, the JWKS endpoint will be queried again on demand. However, the broker polls the URL every sasl.oauthbearer.jwks.endpoint.refresh.ms milliseconds to refresh the cache with any forthcoming keys before any JWT requests that include them are received. If the URL is file-based, the broker will load the JWKS file from a configured location on startup. In the event that the JWT includes a "kid" header value that isn't in the JWKS file, the broker will reject the JWT and authentication will fail.
Type: string Default: null Valid Values: Importance: medium sasl.oauthbearer.token.endpoint.url
The URL for the OAuth/OIDC identity provider. If the URL is HTTP(S)-based, it is the issuer's token endpoint URL to which requests will be made to login based on the configuration in sasl.jaas.config. If the URL is file-based, it specifies a file containing an access token (in JWT serialized form) issued by the OAuth/OIDC identity provider to use for authorization.
Type: string Default: null Valid Values: Importance: medium security.protocol
Protocol used to communicate with brokers. Valid values are: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL.
Type: string Default: PLAINTEXT Valid Values: (case insensitive) [SASL_SSL, PLAINTEXT, SSL, SASL_PLAINTEXT] Importance: medium ssl.enabled.protocols
The list of protocols enabled for SSL connections. The default is 'TLSv1.2,TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' otherwise. With the default value for Java 11, clients and servers will prefer TLSv1.3 if both support it and fallback to TLSv1.2 otherwise (assuming both support at least TLSv1.2). This default should be fine for most cases. Also see the config documentation for `ssl.protocol`.
Type: list Default: TLSv1.2,TLSv1.3 Valid Values: Importance: medium ssl.keystore.type
The file format of the key store file. This is optional for client. The values currently supported by the default `ssl.engine.factory.class` are [JKS, PKCS12, PEM].
Type: string Default: JKS Valid Values: Importance: medium ssl.protocol
The SSL protocol used to generate the SSLContext. The default is 'TLSv1.3' when running with Java 11 or newer, 'TLSv1.2' otherwise. This value should be fine for most use cases. Allowed values in recent JVMs are 'TLSv1.2' and 'TLSv1.3'. 'TLS', 'TLSv1.1', 'SSL', 'SSLv2' and 'SSLv3' may be supported in older JVMs, but their usage is discouraged due to known security vulnerabilities. With the default value for this config and 'ssl.enabled.protocols', clients will downgrade to 'TLSv1.2' if the server does not support 'TLSv1.3'. If this config is set to 'TLSv1.2', clients will not use 'TLSv1.3' even if it is one of the values in ssl.enabled.protocols and the server only supports 'TLSv1.3'.
Type: string Default: TLSv1.3 Valid Values: Importance: medium ssl.provider
The name of the security provider used for SSL connections. Default value is the default security provider of the JVM.
Type: string Default: null Valid Values: Importance: medium ssl.truststore.type
The file format of the trust store file. The values currently supported by the default `ssl.engine.factory.class` are [JKS, PKCS12, PEM].
Type: string Default: JKS Valid Values: Importance: medium admin.timeout.ms
Timeout for administrative tasks, e.g. detecting new topics.
Type: long Default: 60000 (1 minute) Valid Values: Importance: low auto.include.jmx.reporter
Deprecated. Whether to automatically include JmxReporter even if it's not listed in
metric.reporters. This configuration will be removed in Kafka 4.0, users should instead includeorg.apache.kafka.common.metrics.JmxReporterinmetric.reportersin order to enable the JmxReporter.Type: boolean Default: true Valid Values: Importance: low enabled
Whether to replicate source->target.
Type: boolean Default: true Valid Values: Importance: low forwarding.admin.class
Class which extends ForwardingAdmin to define custom cluster resource management (topics, configs, etc). The class must have a constructor with signature
(Mapthat is used to configure a KafkaAdminClient and may also be used to configure clients for external systems if necessary.config) Type: class Default: org.apache.kafka.clients.admin.ForwardingAdmin Valid Values: Importance: low metric.reporters
A list of classes to use as metrics reporters. Implementing the
org.apache.kafka.common.metrics.MetricsReporterinterface allows plugging in classes that will be notified of new metric creation. The JmxReporter is always included to register JMX statistics.Type: list Default: null Valid Values: Importance: low replication.policy.class
Class which defines the remote topic naming convention.
Type: class Default: org.apache.kafka.connect.mirror.DefaultReplicationPolicy Valid Values: Importance: low replication.policy.internal.topic.separator.enabled
Whether to use replication.policy.separator to control the names of topics used for checkpoints and offset syncs. By default, custom separators are used in these topic names; however, if upgrading MirrorMaker 2 from older versions that did not allow for these topic names to be customized, it may be necessary to set this property to 'false' in order to continue using the same names for those topics.
Type: boolean Default: true Valid Values: Importance: low replication.policy.separator
Separator used in remote topic naming convention.
Type: string Default: . Valid Values: Importance: low sasl.kerberos.kinit.cmd
Kerberos kinit command path.
Type: string Default: /usr/bin/kinit Valid Values: Importance: low sasl.kerberos.min.time.before.relogin
Login thread sleep time between refresh attempts.
Type: long Default: 60000 Valid Values: Importance: low sasl.kerberos.ticket.renew.jitter
Percentage of random jitter added to the renewal time.
Type: double Default: 0.05 Valid Values: Importance: low sasl.kerberos.ticket.renew.window.factor
Login thread will sleep until the specified window factor of time from last refresh to ticket's expiry has been reached, at which time it will try to renew the ticket.
Type: double Default: 0.8 Valid Values: Importance: low sasl.login.connect.timeout.ms
The (optional) value in milliseconds for the external authentication provider connection timeout. Currently applies only to OAUTHBEARER.
Type: int Default: null Valid Values: Importance: low sasl.login.read.timeout.ms
The (optional) value in milliseconds for the external authentication provider read timeout. Currently applies only to OAUTHBEARER.
Type: int Default: null Valid Values: Importance: low sasl.login.refresh.buffer.seconds
The amount of buffer time before credential expiration to maintain when refreshing a credential, in seconds. If a refresh would otherwise occur closer to expiration than the number of buffer seconds then the refresh will be moved up to maintain as much of the buffer time as possible. Legal values are between 0 and 3600 (1 hour); a default value of 300 (5 minutes) is used if no value is specified. This value and sasl.login.refresh.min.period.seconds are both ignored if their sum exceeds the remaining lifetime of a credential. Currently applies only to OAUTHBEARER.
Type: short Default: 300 Valid Values: [0,...,3600] Importance: low sasl.login.refresh.min.period.seconds
The desired minimum time for the login refresh thread to wait before refreshing a credential, in seconds. Legal values are between 0 and 900 (15 minutes); a default value of 60 (1 minute) is used if no value is specified. This value and sasl.login.refresh.buffer.seconds are both ignored if their sum exceeds the remaining lifetime of a credential. Currently applies only to OAUTHBEARER.
Type: short Default: 60 Valid Values: [0,...,900] Importance: low sasl.login.refresh.window.factor
Login refresh thread will sleep until the specified window factor relative to the credential's lifetime has been reached, at which time it will try to refresh the credential. Legal values are between 0.5 (50%) and 1.0 (100%) inclusive; a default value of 0.8 (80%) is used if no value is specified. Currently applies only to OAUTHBEARER.
Type: double Default: 0.8 Valid Values: [0.5,...,1.0] Importance: low sasl.login.refresh.window.jitter
The maximum amount of random jitter relative to the credential's lifetime that is added to the login refresh thread's sleep time. Legal values are between 0 and 0.25 (25%) inclusive; a default value of 0.05 (5%) is used if no value is specified. Currently applies only to OAUTHBEARER.
Type: double Default: 0.05 Valid Values: [0.0,...,0.25] Importance: low sasl.login.retry.backoff.max.ms
The (optional) value in milliseconds for the maximum wait between login attempts to the external authentication provider. Login uses an exponential backoff algorithm with an initial wait based on the sasl.login.retry.backoff.ms setting and will double in wait length between attempts up to a maximum wait length specified by the sasl.login.retry.backoff.max.ms setting. Currently applies only to OAUTHBEARER.
Type: long Default: 10000 (10 seconds) Valid Values: Importance: low sasl.login.retry.backoff.ms
The (optional) value in milliseconds for the initial wait between login attempts to the external authentication provider. Login uses an exponential backoff algorithm with an initial wait based on the sasl.login.retry.backoff.ms setting and will double in wait length between attempts up to a maximum wait length specified by the sasl.login.retry.backoff.max.ms setting. Currently applies only to OAUTHBEARER.
Type: long Default: 100 Valid Values: Importance: low sasl.oauthbearer.clock.skew.seconds
The (optional) value in seconds to allow for differences between the time of the OAuth/OIDC identity provider and the broker.
Type: int Default: 30 Valid Values: Importance: low sasl.oauthbearer.expected.audience
The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences. The JWT will be inspected for the standard OAuth "aud" claim and if this value is set, the broker will match the value from JWT's "aud" claim to see if there is an exact match. If there is no match, the broker will reject the JWT and authentication will fail.
Type: list Default: null Valid Values: Importance: low sasl.oauthbearer.expected.issuer
The (optional) setting for the broker to use to verify that the JWT was created by the expected issuer. The JWT will be inspected for the standard OAuth "iss" claim and if this value is set, the broker will match it exactly against what is in the JWT's "iss" claim. If there is no match, the broker will reject the JWT and authentication will fail.
Type: string Default: null Valid Values: Importance: low sasl.oauthbearer.header.urlencode
The (optional) setting to enable the OAuth client to URL-encode the client_id and client_secret in the authorization header in accordance with RFC6749, see here for more details. The default value is set to 'false' for backward compatibility
Type: boolean Default: false Valid Values: Importance: low sasl.oauthbearer.jwks.endpoint.refresh.ms
The (optional) value in milliseconds for the broker to wait between refreshing its JWKS (JSON Web Key Set) cache that contains the keys to verify the signature of the JWT.
Type: long Default: 3600000 (1 hour) Valid Values: Importance: low sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms
The (optional) value in milliseconds for the maximum wait between attempts to retrieve the JWKS (JSON Web Key Set) from the external authentication provider. JWKS retrieval uses an exponential backoff algorithm with an initial wait based on the sasl.oauthbearer.jwks.endpoint.retry.backoff.ms setting and will double in wait length between attempts up to a maximum wait length specified by the sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms setting.
Type: long Default: 10000 (10 seconds) Valid Values: Importance: low sasl.oauthbearer.jwks.endpoint.retry.backoff.ms
The (optional) value in milliseconds for the initial wait between JWKS (JSON Web Key Set) retrieval attempts from the external authentication provider. JWKS retrieval uses an exponential backoff algorithm with an initial wait based on the sasl.oauthbearer.jwks.endpoint.retry.backoff.ms setting and will double in wait length between attempts up to a maximum wait length specified by the sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms setting.
Type: long Default: 100 Valid Values: Importance: low sasl.oauthbearer.scope.claim.name
The OAuth claim for the scope is often named "scope", but this (optional) setting can provide a different name to use for the scope included in the JWT payload's claims if the OAuth/OIDC provider uses a different name for that claim.
Type: string Default: scope Valid Values: Importance: low sasl.oauthbearer.sub.claim.name
The OAuth claim for the subject is often named "sub", but this (optional) setting can provide a different name to use for the subject included in the JWT payload's claims if the OAuth/OIDC provider uses a different name for that claim.
Type: string Default: sub Valid Values: Importance: low ssl.cipher.suites
A list of cipher suites. This is a named combination of authentication, encryption, MAC and key exchange algorithm used to negotiate the security settings for a network connection using TLS or SSL network protocol. By default all the available cipher suites are supported.
Type: list Default: null Valid Values: Importance: low ssl.endpoint.identification.algorithm
The endpoint identification algorithm to validate server hostname using server certificate.
Type: string Default: https Valid Values: Importance: low ssl.engine.factory.class
The class of type org.apache.kafka.common.security.auth.SslEngineFactory to provide SSLEngine objects. Default value is org.apache.kafka.common.security.ssl.DefaultSslEngineFactory. Alternatively, setting this to org.apache.kafka.common.security.ssl.CommonNameLoggingSslEngineFactory will log the common name of expired SSL certificates used by clients to authenticate at any of the brokers with log level INFO. Note that this will cause a tiny delay during establishment of new connections from mTLS clients to brokers due to the extra code for examining the certificate chain provided by the client. Note further that the implementation uses a custom truststore based on the standard Java truststore and thus might be considered a security risk due to not being as mature as the standard one.
Type: class Default: null Valid Values: Importance: low ssl.keymanager.algorithm
The algorithm used by key manager factory for SSL connections. Default value is the key manager factory algorithm configured for the Java Virtual Machine.
Type: string Default: SunX509 Valid Values: Importance: low ssl.secure.random.implementation
The SecureRandom PRNG implementation to use for SSL cryptography operations.
Type: string Default: null Valid Values: Importance: low ssl.trustmanager.algorithm
The algorithm used by trust manager factory for SSL connections. Default value is the trust manager factory algorithm configured for the Java Virtual Machine.
Type: string Default: PKIX Valid Values: Importance: low name
Globally unique name to use for this connector.
Type: string Default: Valid Values: non-empty string without ISO control characters Importance: high connector.class
Name or alias of the class for this connector. Must be a subclass of org.apache.kafka.connect.connector.Connector. If the connector is org.apache.kafka.connect.file.FileStreamSinkConnector, you can either specify this full name, or use "FileStreamSink" or "FileStreamSinkConnector" to make the configuration a bit shorter
Type: string Default: Valid Values: Importance: high tasks.max
Maximum number of tasks to use for this connector.
Type: int Default: 1 Valid Values: [1,...] Importance: high tasks.max.enforce
(Deprecated) Whether to enforce that the tasks.max property is respected by the connector. By default, connectors that generate too many tasks will fail, and existing sets of tasks that exceed the tasks.max property will also be failed. If this property is set to false, then connectors will be allowed to generate more than the maximum number of tasks, and existing sets of tasks that exceed the tasks.max property will be allowed to run. This property is deprecated and will be removed in an upcoming major release.
Type: boolean Default: true Valid Values: Importance: low key.converter
Converter class used to convert between Kafka Connect format and the serialized form that is written to Kafka. This controls the format of the keys in messages written to or read from Kafka, and since this is independent of connectors it allows any connector to work with any serialization format. Examples of common formats include JSON and Avro.
Type: class Default: null Valid Values: A concrete subclass of org.apache.kafka.connect.storage.Converter, A class with a public, no-argument constructor Importance: low value.converter
Converter class used to convert between Kafka Connect format and the serialized form that is written to Kafka. This controls the format of the values in messages written to or read from Kafka, and since this is independent of connectors it allows any connector to work with any serialization format. Examples of common formats include JSON and Avro.
Type: class Default: null Valid Values: A concrete subclass of org.apache.kafka.connect.storage.Converter, A class with a public, no-argument constructor Importance: low header.converter
HeaderConverter class used to convert between Kafka Connect format and the serialized form that is written to Kafka. This controls the format of the header values in messages written to or read from Kafka, and since this is independent of connectors it allows any connector to work with any serialization format. Examples of common formats include JSON and Avro. By default, the SimpleHeaderConverter is used to serialize header values to strings and deserialize them by inferring the schemas.
Type: class Default: null Valid Values: A concrete subclass of org.apache.kafka.connect.storage.HeaderConverter, A class with a public, no-argument constructor Importance: low config.action.reload
The action that Connect should take on the connector when changes in external configuration providers result in a change in the connector's configuration properties. A value of 'none' indicates that Connect will do nothing. A value of 'restart' indicates that Connect should restart/reload the connector with the updated configuration properties.The restart may actually be scheduled in the future if the external configuration provider indicates that a configuration value will expire in the future.
Type: string Default: restart Valid Values: [none, restart] Importance: low transforms
Aliases for the transformations to be applied to records.
Type: list Default: "" Valid Values: non-null string, unique transformation aliases Importance: low predicates
Aliases for the predicates used by transformations.
Type: list Default: "" Valid Values: non-null string, unique predicate aliases Importance: low errors.retry.timeout
The maximum duration in milliseconds that a failed operation will be reattempted. The default is 0, which means no retries will be attempted. Use -1 for infinite retries.
Type: long Default: 0 Valid Values: Importance: medium errors.retry.delay.max.ms
The maximum duration in milliseconds between consecutive retry attempts. Jitter will be added to the delay once this limit is reached to prevent thundering herd issues.
Type: long Default: 60000 (1 minute) Valid Values: Importance: medium errors.tolerance
Behavior for tolerating errors during connector operation. 'none' is the default value and signals that any error will result in an immediate connector task failure; 'all' changes the behavior to skip over problematic records.
Type: string Default: none Valid Values: [none, all] Importance: medium errors.log.enable
If true, write each error and the details of the failed operation and problematic record to the Connect application log. This is 'false' by default, so that only errors that are not tolerated are reported.
Type: boolean Default: false Valid Values: Importance: medium errors.log.include.messages
Whether to include in the log the Connect record that resulted in a failure. For sink records, the topic, partition, offset, and timestamp will be logged. For source records, the key and value (and their schemas), all headers, and the timestamp, Kafka topic, Kafka partition, source partition, and source offset will be logged. This is 'false' by default, which will prevent record keys, values, and headers from being written to log files.
Type: boolean Default: false Valid Values: Importance: medium
MirrorMaker Source Configs
Below is the configuration of MirrorMaker 2 source connector for replicating topics.
config.properties.blacklist
Deprecated. Use config.properties.exclude instead.
Type: list Default: null Valid Values: Importance: high config.properties.exclude
Topic config properties that should not be replicated. Supports comma-separated property names and regexes.
Type: list Default: follower\.replication\.throttled\.replicas,leader\.replication\.throttled\.replicas,message\.timestamp\.difference\.max\.ms,message\.timestamp\.type,unclean\.leader\.election\.enable,min\.insync\.replicas Valid Values: Importance: high topics
Topics to replicate. Supports comma-separated topic names and regexes.
Type: list Default: .* Valid Values: Importance: high topics.blacklist
Deprecated. Use topics.exclude instead.
Type: list Default: null Valid Values: Importance: high topics.exclude
Excluded topics. Supports comma-separated topic names and regexes. Excludes take precedence over includes.
Type: list Default: .*[\-\.]internal,.*\.replica,__.* Valid Values: Importance: high add.source.alias.to.metrics
Deprecated. Whether to tag metrics with the source cluster alias. Metrics have the target, topic and partition tags. When this setting is enabled, it adds the source tag. This configuration will be removed in Kafka 4.0 and the default behavior will be to always have the source tag.
Type: boolean Default: false Valid Values: Importance: low config.property.filter.class
ConfigPropertyFilter to use. Selects topic config properties to replicate.
Type: class Default: org.apache.kafka.connect.mirror.DefaultConfigPropertyFilter Valid Values: Importance: low consumer.poll.timeout.ms
Timeout when polling source cluster.
Type: long Default: 1000 (1 second) Valid Values: Importance: low emit.offset-syncs.enabled
Whether to store the new offset of the replicated records in offset-syncs topic or not. MirrorCheckpointConnector will not be able to sync group offsets or emit checkpoints if emit.checkpoints.enabled and/or sync.group.offsets.enabled are enabled while emit.offset-syncs.enabled is disabled.
Type: boolean Default: true Valid Values: Importance: low offset-syncs.topic.location
The location (source/target) of the offset-syncs topic.
Type: string Default: source Valid Values: [source, target] Importance: low offset-syncs.topic.replication.factor
Replication factor for offset-syncs topic.
Type: short Default: 3 Valid Values: Importance: low offset.lag.max
How out-of-sync a remote partition can be before it is resynced.
Type: long Default: 100 Valid Values: Importance: low refresh.topics.enabled
Whether to periodically check for new topics and partitions.
Type: boolean Default: true Valid Values: Importance: low refresh.topics.interval.seconds
Frequency of topic refresh.
Type: long Default: 600 Valid Values: Importance: low replication.factor
Replication factor for newly created remote topics.
Type: int Default: 2 Valid Values: Importance: low sync.topic.acls.enabled
Whether to periodically configure remote topic ACLs to match their corresponding upstream topics.
Type: boolean Default: true Valid Values: Importance: low sync.topic.acls.interval.seconds
Frequency of topic ACL sync.
Type: long Default: 600 Valid Values: Importance: low sync.topic.configs.enabled
Whether to periodically configure remote topics to match their corresponding upstream topics.
Type: boolean Default: true Valid Values: Importance: low sync.topic.configs.interval.seconds
Frequency of topic config sync.
Type: long Default: 600 Valid Values: Importance: low topic.filter.class
TopicFilter to use. Selects topics to replicate.
Type: class Default: org.apache.kafka.connect.mirror.DefaultTopicFilter Valid Values: Importance: low use.incremental.alter.configs
Deprecated. Which API to use for syncing topic configs. The valid values are 'requested', 'required' and 'never'. By default, set to 'requested', which means the IncrementalAlterConfigs API is being used for syncing topic configurations and if any request receives an error from an incompatible broker, it will fallback to using the deprecated AlterConfigs API. If explicitly set to 'required', the IncrementalAlterConfigs API is used without the fallback logic and +if it receives an error from an incompatible broker, the connector will fail.If explicitly set to 'never', the AlterConfig is always used.This setting will be removed and the behaviour of 'required' will be used in Kafka 4.0, therefore users should ensure that target broker is at least 2.3.0
Type: string Default: requested Valid Values: [requested, required, never] Importance: low
MirrorMaker Checkpoint Configs
Below is the configuration of MirrorMaker 2 checkpoint connector for emitting consumer offset checkpoints.
groups
Consumer groups to replicate. Supports comma-separated group IDs and regexes.
Type: list Default: .* Valid Values: Importance: high groups.blacklist
Deprecated. Use groups.exclude instead.
Type: list Default: null Valid Values: Importance: high groups.exclude
Exclude groups. Supports comma-separated group IDs and regexes. Excludes take precedence over includes.
Type: list Default: console-consumer-.*,connect-.*,__.* Valid Values: Importance: high checkpoints.topic.replication.factor
Replication factor for checkpoints topic.
Type: short Default: 3 Valid Values: Importance: low consumer.poll.timeout.ms
Timeout when polling source cluster.
Type: long Default: 1000 (1 second) Valid Values: Importance: low emit.checkpoints.enabled
Whether to replicate consumer offsets to target cluster.
Type: boolean Default: true Valid Values: Importance: low emit.checkpoints.interval.seconds
Frequency of checkpoints.
Type: long Default: 60 Valid Values: Importance: low group.filter.class
GroupFilter to use. Selects consumer groups to replicate.
Type: class Default: org.apache.kafka.connect.mirror.DefaultGroupFilter Valid Values: Importance: low offset-syncs.topic.location
The location (source/target) of the offset-syncs topic.
Type: string Default: source Valid Values: [source, target] Importance: low refresh.groups.enabled
Whether to periodically check for new consumer groups.
Type: boolean Default: true Valid Values: Importance: low refresh.groups.interval.seconds
Frequency of group refresh.
Type: long Default: 600 Valid Values: Importance: low sync.group.offsets.enabled
Whether to periodically write the translated offsets to __consumer_offsets topic in target cluster, as long as no active consumers in that group are connected to the target cluster
Type: boolean Default: false Valid Values: Importance: low sync.group.offsets.interval.seconds
Frequency of consumer group offset sync.
Type: long Default: 60 Valid Values: Importance: low topic.filter.class
TopicFilter to use. Selects topics to replicate.
Type: class Default: org.apache.kafka.connect.mirror.DefaultTopicFilter Valid Values: Importance: low
MirrorMaker HeartBeat Configs
Below is the configuration of MirrorMaker 2 heartbeat connector for checking connectivity between connectors and clusters.
emit.heartbeats.enabled
Whether to emit heartbeats to target cluster.
Type: boolean Default: true Valid Values: Importance: low emit.heartbeats.interval.seconds
Frequency of heartbeats.
Type: long Default: 1 Valid Values: Importance: low heartbeats.topic.replication.factor
Replication factor for heartbeats topic.
Type: short Default: 3 Valid Values: Importance: low