Created
September 12, 2018 16:21
-
-
Save srijiths/c5516672ffcbd146d2afdb2a1dfbed05 to your computer and use it in GitHub Desktop.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
## | |
# Licensed to the Apache Software Foundation (ASF) under one or more | |
# contributor license agreements. See the NOTICE file distributed with | |
# this work for additional information regarding copyright ownership. | |
# The ASF licenses this file to You under the Apache License, Version 2.0 | |
# (the "License"); you may not use this file except in compliance with | |
# the License. You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
## | |
# This file contains some of the configurations for the Kafka Connect distributed worker. This file is intended | |
# to be used with the examples, and some settings may differ from those used in a production system, especially | |
# the `bootstrap.servers` and those specifying replication factors. | |
# A list of host/port pairs to use for establishing the initial connection to the Kafka cluster. | |
bootstrap.servers=XX.XX.XX.XX:9092,XX.XX.XX.XX:9092 | |
# unique name for the cluster, used in forming the Connect cluster group. Note that this must not conflict with consumer group IDs | |
group.id=connect-cluster | |
# The converters specify the format of data in Kafka and how to translate it into Connect data. Every Connect user will | |
# need to configure these based on the format they want their data in when loaded from or stored into Kafka | |
#key.converter=org.apache.kafka.connect.json.JsonConverter | |
#value.converter=org.apache.kafka.connect.json.JsonConverter | |
key.converter=org.apache.kafka.connect.storage.StringConverter | |
value.converter=org.apache.kafka.connect.storage.StringConverter | |
# Converter-specific settings can be passed in by prefixing the Converter's setting with the converter we want to apply | |
# it to | |
#key.converter.schemas.enable=true | |
#value.converter.schemas.enable=true | |
key.converter.schemas.enable=false | |
value.converter.schemas.enable=false | |
# The internal converter used for offsets, config, and status data is configurable and must be specified, but most users will | |
# always want to use the built-in default. Offset, config, and status data is never visible outside of Kafka Connect in this format. | |
internal.key.converter=org.apache.kafka.connect.json.JsonConverter | |
internal.value.converter=org.apache.kafka.connect.json.JsonConverter | |
internal.key.converter.schemas.enable=false | |
internal.value.converter.schemas.enable=false | |
# Topic to use for storing offsets. This topic should have many partitions and be replicated and compacted. | |
# Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create | |
# the topic before starting Kafka Connect if a specific topic configuration is needed. | |
# Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. | |
# Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able | |
# to run this example on a single-broker cluster and so here we instead set the replication factor to 1. | |
offset.storage.topic=connect-offsets | |
offset.storage.replication.factor=1 | |
#offset.storage.partitions=25 | |
# Topic to use for storing connector and task configurations; note that this should be a single partition, highly replicated, | |
# and compacted topic. Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create | |
# the topic before starting Kafka Connect if a specific topic configuration is needed. | |
# Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. | |
# Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able | |
# to run this example on a single-broker cluster and so here we instead set the replication factor to 1. | |
config.storage.topic=connect-configs | |
config.storage.replication.factor=1 | |
# Topic to use for storing statuses. This topic can have multiple partitions and should be replicated and compacted. | |
# Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create | |
# the topic before starting Kafka Connect if a specific topic configuration is needed. | |
# Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. | |
# Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able | |
# to run this example on a single-broker cluster and so here we instead set the replication factor to 1. | |
status.storage.topic=connect-status | |
status.storage.replication.factor=1 | |
#status.storage.partitions=5 | |
# Flush much faster than normal, which is useful for testing/debugging | |
offset.flush.interval.ms=10000 | |
# These are provided to inform the user about the presence of the REST host and port configs | |
# Hostname & Port for the REST API to listen on. If this is set, it will bind to the interface used to listen to requests. | |
#rest.host.name= | |
#rest.port=8083 | |
# The Hostname & Port that will be given out to other workers to connect to i.e. URLs that are routable from other servers. | |
#rest.advertised.host.name= | |
#rest.advertised.port= | |
# Set to a list of filesystem paths separated by commas (,) to enable class loading isolation for plugins | |
# (connectors, converters, transformations). The list should consist of top level directories that include | |
# any combination of: | |
# a) directories immediately containing jars with plugins and their dependencies | |
# b) uber-jars with plugins and their dependencies | |
# c) directories immediately containing the package directory structure of classes of plugins and their dependencies | |
# Examples: | |
# plugin.path=/usr/local/share/java,/usr/local/share/kafka/plugins,/opt/connectors, | |
plugin.path=/usr/share/java |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment