./kafka-server-start.sh ../config/server.properties [2018-09-06 08:00:50,719] INFO KafkaConfig values: advertised.host.name = null advertised.listeners = SASL_PLAINTEXT://kafka1.hostname.com:9092 advertised.port = null authorizer.class.name = auto.create.topics.enable = true auto.leader.rebalance.enable = true background.threads = 10 broker.id = 1 broker.id.generation.enable = true broker.rack = null compression.type = gzip connections.max.idle.ms = 600000 controlled.shutdown.enable = true controlled.shutdown.max.retries = 3 controlled.shutdown.retry.backoff.ms = 5000 controller.socket.timeout.ms = 30000 create.topic.policy.class.name = null default.replication.factor = 2 delete.topic.enable = false fetch.purgatory.purge.interval.requests = 1000 group.max.session.timeout.ms = 300000 group.min.session.timeout.ms = 6000 host.name = kafka1.hostname.com inter.broker.listener.name = null inter.broker.protocol.version = 0.10.2-IV0 leader.imbalance.check.interval.seconds = 300 leader.imbalance.per.broker.percentage = 10 listener.security.protocol.map = SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,TRACE:TRACE,SASL_SSL:SASL_SSL,PLAINTEXT:PLAINTEXT listeners = SASL_PLAINTEXT://kafka1.hostname.com:9092 log.cleaner.backoff.ms = 15000 log.cleaner.dedupe.buffer.size = 134217728 log.cleaner.delete.retention.ms = 86400000 log.cleaner.enable = true log.cleaner.io.buffer.load.factor = 0.9 log.cleaner.io.buffer.size = 524288 log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 log.cleaner.min.cleanable.ratio = 0.5 log.cleaner.min.compaction.lag.ms = 0 log.cleaner.threads = 1 log.cleanup.policy = [delete] log.dir = /tmp/kafka-logs log.dirs = /home/deepak/kafka/kafka-logs log.flush.interval.messages = 9223372036854775807 log.flush.interval.ms = null log.flush.offset.checkpoint.interval.ms = 60000 log.flush.scheduler.interval.ms = 9223372036854775807 log.index.interval.bytes = 4096 log.index.size.max.bytes = 10485760 log.message.format.version = 0.10.2-IV0 log.message.timestamp.difference.max.ms = 9223372036854775807 log.message.timestamp.type = CreateTime log.preallocate = false log.retention.bytes = -1 log.retention.check.interval.ms = 300000 log.retention.hours = 168 log.retention.minutes = null log.retention.ms = null log.roll.hours = 168 log.roll.jitter.hours = 0 log.roll.jitter.ms = null log.roll.ms = null log.segment.bytes = 1073741824 log.segment.delete.delay.ms = 60000 max.connections.per.ip = 2147483647 max.connections.per.ip.overrides = message.max.bytes = 40000000 metric.reporters = [] metrics.num.samples = 2 metrics.recording.level = INFO metrics.sample.window.ms = 30000 min.insync.replicas = 1 num.io.threads = 8 num.network.threads = 9 num.partitions = 2 num.recovery.threads.per.data.dir = 1 num.replica.fetchers = 1 offset.metadata.max.bytes = 4096 offsets.commit.required.acks = -1 offsets.commit.timeout.ms = 5000 offsets.load.buffer.size = 5242880 offsets.retention.check.interval.ms = 600000 offsets.retention.minutes = 1440 offsets.topic.compression.codec = 0 offsets.topic.num.partitions = 50 offsets.topic.replication.factor = 3 offsets.topic.segment.bytes = 104857600 port = 9092 principal.builder.class = class org.apache.kafka.common.security.auth.DefaultPrincipalBuilder producer.purgatory.purge.interval.requests = 1000 queued.max.requests = 500 quota.consumer.default = 9223372036854775807 quota.producer.default = 9223372036854775807 quota.window.num = 11 quota.window.size.seconds = 1 replica.fetch.backoff.ms = 1000 replica.fetch.max.bytes = 104857600 replica.fetch.min.bytes = 1 replica.fetch.response.max.bytes = 10485760 replica.fetch.wait.max.ms = 500 replica.high.watermark.checkpoint.interval.ms = 5000 replica.lag.time.max.ms = 10000 replica.socket.receive.buffer.bytes = 65536 replica.socket.timeout.ms = 30000 replication.quota.window.num = 11 replication.quota.window.size.seconds = 1 request.timeout.ms = 30000 reserved.broker.max.id = 1000 sasl.enabled.mechanisms = [GSSAPI] sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.min.time.before.relogin = 60000 sasl.kerberos.principal.to.local.rules = [DEFAULT] sasl.kerberos.service.name = kafka sasl.kerberos.ticket.renew.jitter = 0.05 sasl.kerberos.ticket.renew.window.factor = 0.8 sasl.mechanism.inter.broker.protocol = GSSAPI security.inter.broker.protocol = SASL_PLAINTEXT socket.receive.buffer.bytes = 102400 socket.request.max.bytes = 104857600 socket.send.buffer.bytes = 102400 ssl.cipher.suites = null ssl.client.auth = none ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.endpoint.identification.algorithm = null ssl.key.password = null ssl.keymanager.algorithm = SunX509 ssl.keystore.location = null ssl.keystore.password = null ssl.keystore.type = JKS ssl.protocol = TLS ssl.provider = null ssl.secure.random.implementation = null ssl.trustmanager.algorithm = PKIX ssl.truststore.location = null ssl.truststore.password = null ssl.truststore.type = JKS unclean.leader.election.enable = true zookeeper.connect = kafka1.hostname.com:2182 zookeeper.connection.timeout.ms = 30000 zookeeper.session.timeout.ms = 30000 zookeeper.set.acl = false zookeeper.sync.time.ms = 2000 (kafka.server.KafkaConfig) [2018-09-06 08:00:50,764] INFO starting (kafka.server.KafkaServer) [2018-09-06 08:00:50,765] INFO Connecting to zookeeper on kafka1.hostname.com:2182 (kafka.server.KafkaServer) [2018-09-06 08:00:50,774] INFO JAAS File name: /home/deepak/kafka/kafka_jaas.conf (org.I0Itec.zkclient.ZkClient) [2018-09-06 08:00:50,775] INFO Starting ZkClient event thread. (org.I0Itec.zkclient.ZkEventThread) [2018-09-06 08:00:50,778] INFO Client environment:zookeeper.version=3.4.9-1757313, built on 08/23/2016 06:50 GMT (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:host.name=kafka1.hostname.com (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.version=1.8.0_171 (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.vendor=Oracle Corporation (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.home=/usr/lib/jvm/java-8-openjdk-amd64/jre (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.class.path=:/home/deepak/kafka/bin/../libs/aopalliance-repackaged-2.5.0-b05.jar:/home/deepak/kafka/bin/../libs/argparse4j-0.7.0.jar:/home/deepak/kafka/bin/../libs/connect-api-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/connect-file-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/connect-json-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/connect-runtime-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/connect-transforms-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/guava-18.0.jar:/home/deepak/kafka/bin/../libs/hk2-api-2.5.0-b05.jar:/home/deepak/kafka/bin/../libs/hk2-locator-2.5.0-b05.jar:/home/deepak/kafka/bin/../libs/hk2-utils-2.5.0-b05.jar:/home/deepak/kafka/bin/../libs/jackson-annotations-2.8.0.jar:/home/deepak/kafka/bin/../libs/jackson-annotations-2.8.5.jar:/home/deepak/kafka/bin/../libs/jackson-core-2.8.5.jar:/home/deepak/kafka/bin/../libs/jackson-databind-2.8.5.jar:/home/deepak/kafka/bin/../libs/jackson-jaxrs-base-2.8.5.jar:/home/deepak/kafka/bin/../libs/jackson-jaxrs-json-provider-2.8.5.jar:/home/deepak/kafka/bin/../libs/jackson-module-jaxb-annotations-2.8.5.jar:/home/deepak/kafka/bin/../libs/javassist-3.20.0-GA.jar:/home/deepak/kafka/bin/../libs/javax.annotation-api-1.2.jar:/home/deepak/kafka/bin/../libs/javax.inject-1.jar:/home/deepak/kafka/bin/../libs/javax.inject-2.5.0-b05.jar:/home/deepak/kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/home/deepak/kafka/bin/../libs/javax.ws.rs-api-2.0.1.jar:/home/deepak/kafka/bin/../libs/jersey-client-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-common-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-container-servlet-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-container-servlet-core-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-guava-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-media-jaxb-2.24.jar:/home/deepak/kafka/bin/../libs/jersey-server-2.24.jar:/home/deepak/kafka/bin/../libs/jetty-continuation-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-http-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-io-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-security-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-server-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-servlet-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-servlets-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jetty-util-9.2.15.v20160210.jar:/home/deepak/kafka/bin/../libs/jopt-simple-5.0.3.jar:/home/deepak/kafka/bin/../libs/kafka_2.12-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/kafka_2.12-0.10.2.0-sources.jar:/home/deepak/kafka/bin/../libs/kafka_2.12-0.10.2.0-test-sources.jar:/home/deepak/kafka/bin/../libs/kafka-clients-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/kafka-log4j-appender-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/kafka-streams-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/kafka-streams-examples-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/kafka-tools-0.10.2.0.jar:/home/deepak/kafka/bin/../libs/log4j-1.2.17.jar:/home/deepak/kafka/bin/../libs/lz4-1.3.0.jar:/home/deepak/kafka/bin/../libs/metrics-core-2.2.0.jar:/home/deepak/kafka/bin/../libs/osgi-resource-locator-1.0.1.jar:/home/deepak/kafka/bin/../libs/reflections-0.9.10.jar:/home/deepak/kafka/bin/../libs/rocksdbjni-5.0.1.jar:/home/deepak/kafka/bin/../libs/scala-library-2.12.1.jar:/home/deepak/kafka/bin/../libs/scala-parser-combinators_2.12-1.0.4.jar:/home/deepak/kafka/bin/../libs/slf4j-api-1.7.21.jar:/home/deepak/kafka/bin/../libs/slf4j-log4j12-1.7.21.jar:/home/deepak/kafka/bin/../libs/snappy-java-1.1.2.6.jar:/home/deepak/kafka/bin/../libs/validation-api-1.1.0.Final.jar:/home/deepak/kafka/bin/../libs/zkclient-0.10.jar:/home/deepak/kafka/bin/../libs/zookeeper-3.4.9.jar (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib/x86_64-linux-gnu/jni:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/jni:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,778] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:os.version=4.4.0-128-generic (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:user.name=root (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:user.home=/root (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Client environment:user.dir=/home/deepak/kafka/bin (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,779] INFO Initiating client connection, connectString=kafka1.hostname.com:2182 sessionTimeout=30000 watcher=org.I0Itec.zkclient.ZkClient@56de5251 (org.apache.zookeeper.ZooKeeper) [2018-09-06 08:00:50,787] INFO Waiting for keeper state SaslAuthenticated (org.I0Itec.zkclient.ZkClient) Debug is true storeKey true useTicketCache false useKeyTab true doNotPrompt false ticketCache is null isInitiator true KeyTab is /tmp/kafka.keytab refreshKrb5Config is true principal is kafka/kafka1.hostname.com@MSTORM.COM tryFirstPass is false useFirstPass is false storePass is false clearPass is false Refreshing Kerberos configuration Java config name: /etc/krb5.conf Loaded from Java config >>> KdcAccessibility: reset >>> KdcAccessibility: reset >>> KeyTabInputStream, readName(): MSTORM.COM >>> KeyTabInputStream, readName(): kafka >>> KeyTabInputStream, readName(): kafka1.hostname.com >>> KeyTab: load() entry length: 87; type: 18 >>> KeyTabInputStream, readName(): MSTORM.COM >>> KeyTabInputStream, readName(): kafka >>> KeyTabInputStream, readName(): kafka1.hostname.com >>> KeyTab: load() entry length: 71; type: 23 >>> KeyTabInputStream, readName(): MSTORM.COM >>> KeyTabInputStream, readName(): kafka >>> KeyTabInputStream, readName(): kafka1.hostname.com >>> KeyTab: load() entry length: 79; type: 16 >>> KeyTabInputStream, readName(): MSTORM.COM >>> KeyTabInputStream, readName(): kafka >>> KeyTabInputStream, readName(): kafka1.hostname.com >>> KeyTab: load() entry length: 63; type: 1 Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. >>> KrbAsReq creating message >>> KrbKdcReq send: kdc=kerberosServer.hostname.com UDP:88, timeout=30000, number of retries =3, #bytes=159 >>> KDCCommunication: kdc=kerberosServer.hostname.com UDP:88, timeout=30000,Attempt =1, #bytes=159 >>> KrbKdcReq send: #bytes read=387 >>>Pre-Authentication Data: PA-DATA type = 136 >>>Pre-Authentication Data: PA-DATA type = 19 PA-ETYPE-INFO2 etype = 18, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 23, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 16, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null >>>Pre-Authentication Data: PA-DATA type = 2 PA-ENC-TIMESTAMP >>>Pre-Authentication Data: PA-DATA type = 133 >>> KdcAccessibility: remove kerberosServer.hostname.com >>> KDCRep: init() encoding tag is 126 req type is 11 >>>KRBError: cTime is Sun Mar 28 10:13:01 EST 1999 922633981000 sTime is Thu Sep 06 08:00:50 EDT 2018 1536235250000 suSec is 806560 error code is 25 error Message is Additional pre-authentication required cname is kafka/kafka1.hostname.com@MSTORM.COM sname is krbtgt/MSTORM.COM@MSTORM.COM eData provided. msgType is 30 >>>Pre-Authentication Data: PA-DATA type = 136 >>>Pre-Authentication Data: PA-DATA type = 19 PA-ETYPE-INFO2 etype = 18, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 23, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 16, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null >>>Pre-Authentication Data: PA-DATA type = 2 PA-ENC-TIMESTAMP >>>Pre-Authentication Data: PA-DATA type = 133 KRBError received: NEEDED_PREAUTH KrbAsReqBuilder: PREAUTH FAILED/REQ, re-send AS-REQ Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbAsReq creating message >>> KrbKdcReq send: kdc=kerberosServer.hostname.com UDP:88, timeout=30000, number of retries =3, #bytes=246 >>> KDCCommunication: kdc=kerberosServer.hostname.com UDP:88, timeout=30000,Attempt =1, #bytes=246 >>> KrbKdcReq send: #bytes read=707 >>> KdcAccessibility: remove kerberosServer.hostname.com Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbAsRep cons in KrbAsReq.getReply kafka/kafka1.hostname.com principal is kafka/kafka1.hostname.com@MSTORM.COM Will use keytab Commit Succeeded [2018-09-06 08:00:50,875] INFO successfully logged in. (org.apache.zookeeper.Login) [2018-09-06 08:00:50,876] INFO TGT refresh thread started. (org.apache.zookeeper.Login) [2018-09-06 08:00:50,876] INFO TGT valid starting at: Thu Sep 06 08:00:50 EDT 2018 (org.apache.zookeeper.Login) [2018-09-06 08:00:50,876] INFO TGT expires: Thu Sep 06 18:00:50 EDT 2018 (org.apache.zookeeper.Login) [2018-09-06 08:00:50,877] INFO TGT refresh sleeping until: Thu Sep 06 16:25:56 EDT 2018 (org.apache.zookeeper.Login) [2018-09-06 08:00:50,879] INFO Client will use GSSAPI as SASL mechanism. (org.apache.zookeeper.client.ZooKeeperSaslClient) [2018-09-06 08:00:50,883] INFO Opening socket connection to server kafka1.hostname.com/10.13.10.26:2182. Will attempt to SASL-authenticate using Login Context section 'Client' (org.apache.zookeeper.ClientCnxn) [2018-09-06 08:00:50,886] INFO Socket connection established to kafka1.hostname.com/10.13.10.26:2182, initiating session (org.apache.zookeeper.ClientCnxn) [2018-09-06 08:00:50,994] INFO Session establishment complete on server kafka1.hostname.com/10.13.10.26:2182, sessionid = 0x165aec0a3eb0000, negotiated timeout = 30000 (org.apache.zookeeper.ClientCnxn) [2018-09-06 08:00:50,997] INFO zookeeper state changed (SyncConnected) (org.I0Itec.zkclient.ZkClient) Found ticket for kafka/kafka1.hostname.com@MSTORM.COM to go to krbtgt/MSTORM.COM@MSTORM.COM expiring on Thu Sep 06 18:00:50 EDT 2018 Entered Krb5Context.initSecContext with state=STATE_NEW Found ticket for kafka/kafka1.hostname.com@MSTORM.COM to go to krbtgt/MSTORM.COM@MSTORM.COM expiring on Thu Sep 06 18:00:50 EDT 2018 Service ticket not found in the subject >>> Credentials acquireServiceCreds: same realm Using builtin default etypes for default_tgs_enctypes default etypes for default_tgs_enctypes: 18 17 16 23. >>> CksumType: sun.security.krb5.internal.crypto.RsaMd5CksumType >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbKdcReq send: kdc=kerberosServer.hostname.com UDP:88, timeout=30000, number of retries =3, #bytes=692 >>> KDCCommunication: kdc=kerberosServer.hostname.com UDP:88, timeout=30000,Attempt =1, #bytes=692 >>> KrbKdcReq send: #bytes read=709 >>> KdcAccessibility: remove kerberosServer.hostname.com >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbApReq: APOptions are 00000000 00000000 00000000 00000000 >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType Krb5Context setting mySeqNumber to: 1025024537 Krb5Context setting peerSeqNumber to: 0 Created InitSecContextToken: 0000: 01 00 6E 82 02 73 30 82 02 6F A0 03 02 01 05 A1 ..n..s0..o...... 0010: 03 02 01 0E A2 07 03 05 00 00 00 00 00 A3 82 01 ................ 0020: 74 61 82 01 70 30 82 01 6C A0 03 02 01 05 A1 0C ta..p0..l....... 0030: 1B 0A 4D 53 54 4F 52 4D 2E 43 4F 4D A2 2B 30 29 ..MSTORM.COM.+0) 0040: A0 03 02 01 00 A1 22 30 20 1B 09 7A 6F 6F 6B 65 ......"0 ..zooke 0050: 65 70 65 72 1B 13 75 62 75 6E 74 75 32 36 2E 6D eper..ubuntu26.m 0060: 73 74 6F 72 6D 2E 63 6F 6D A3 82 01 28 30 82 01 storm.com...(0.. 0070: 24 A0 03 02 01 12 A1 03 02 01 02 A2 82 01 16 04 $............... 0080: 82 01 12 9C 6E 69 07 04 E9 ED FE 1E A3 EC 85 A4 ....ni.......... 0090: 79 F8 57 85 31 CE 1B 54 25 57 60 A9 80 92 77 04 y.W.1..T%W`...w. 00A0: 0A B5 39 D4 35 CF 2F CE C1 CB 69 28 4F 9F ED 83 ..9.5./...i(O... 00B0: B8 0B 58 1D BA 95 DE C5 C6 48 A1 5D 1E DD 4E 07 ..X......H.]..N. 00C0: 4F 7C DF C6 E1 10 4D 18 B1 24 21 45 21 86 E3 F1 O.....M..$!E!... 00D0: 5B 06 61 29 FF 51 6C 99 50 79 39 6E 81 A7 9F 6A [.a).Ql.Py9n...j 00E0: C1 FD 56 5C 9F 78 64 E3 FC F3 7C 9B C9 52 ED 94 ..V\.xd......R.. 00F0: 70 65 3F EB 31 FB 2E 74 25 BF 5F 1D BC C6 B9 A8 pe?.1..t%._..... 0100: E3 22 3B 9E B1 E7 42 6D 62 22 70 90 A1 75 A4 01 .";...Bmb"p..u.. 0110: 8C 19 39 72 E7 DD 64 45 09 A3 9A E6 F6 F2 F2 AE ..9r..dE........ 0120: F9 02 FF DB 37 39 1A B8 C5 C8 7A E4 72 8E 5D 24 ....79....z.r.]$ 0130: CA 6C ED F9 F5 1B B9 2F 7D 4C DE 70 BA D3 18 2E .l...../.L.p.... 0140: 98 C4 E7 00 B4 D8 7E 87 4E CD 48 3A 4E 80 8D E1 ........N.H:N... 0150: F3 3B 0A 29 6C 94 A7 57 81 B1 0A B7 FC 17 F3 D9 .;.)l..W........ 0160: 4F AD BD 8E DB 72 1B 4F B5 A7 B1 57 F4 08 FF 42 O....r.O...W...B 0170: ED 5B EF 7C D9 4A 27 AE 43 53 C1 89 8D D0 6B BD .[...J'.CS....k. 0180: 86 93 09 06 1A EC 0A 54 C0 D2 22 93 B1 8E 95 C3 .......T.."..... 0190: 99 EF C3 7E 53 A4 81 E1 30 81 DE A0 03 02 01 12 ....S...0....... 01A0: A2 81 D6 04 81 D3 A4 F7 AA 44 88 C0 A3 F0 0E 8B .........D...... 01B0: B1 83 7A DD 8E E2 D3 5D 05 D4 63 7E DD A3 1E F6 ..z....]..c..... 01C0: 24 77 24 E5 08 4B 7A BC 8F BD 00 D5 E2 A3 29 61 $w$..Kz.......)a 01D0: AC 50 D1 0F 1C 73 F5 DE 09 19 D4 F4 F8 3F 74 54 .P...s.......?tT 01E0: 6E 19 FF AD E3 30 11 43 62 EC B1 7D D3 F7 2E 56 n....0.Cb......V 01F0: 17 88 A0 EF DA E3 AB D3 87 8B 5C A3 BE 5C 9B ED ..........\..\.. 0200: 3B 28 78 A3 D7 E5 D4 F8 89 3E B8 C8 FE EA 11 E9 ;(x......>...... 0210: A0 16 ED 63 31 9C 73 05 E3 D7 7E 12 11 CD C4 F8 ...c1.s......... 0220: 8D 06 D4 AD F3 A0 78 3F 03 4A F8 42 83 60 52 86 ......x?.J.B.`R. 0230: 7C 76 3F 11 4C 05 DF 24 E4 49 37 48 42 FE C6 32 .v?.L..$.I7HB..2 0240: 20 CB 14 30 B3 58 11 92 B1 CA C0 66 B1 F8 23 A3 ..0.X.....f..#. 0250: FA 94 A8 8C C6 9A 25 E3 24 E4 9F A3 B0 CB D9 C0 ......%.$....... 0260: 89 F9 61 1C EB 7E BE 88 52 36 96 F1 5A 24 03 6F ..a.....R6..Z$.o 0270: BD 01 A9 97 70 9E B7 5D F4 ....p..]. Krb5Context.unwrap: token=[05 04 01 ff 00 0c 00 00 00 00 00 00 3d 18 a2 19 01 01 00 00 b6 d5 71 73 c4 07 df 63 c0 09 bd 50 ] Krb5Context.unwrap: data=[01 01 00 00 ] Krb5Context.wrap: data=[01 01 00 00 6b 61 66 6b 61 2f 75 62 75 6e 74 75 32 36 2e 6d 73 74 6f 72 6d 2e 63 6f 6d 40 4d 53 54 4f 52 4d 2e 43 4f 4d ] Krb5Context.wrap: token=[05 04 00 ff 00 0c 00 00 00 00 00 00 3d 18 a2 19 01 01 00 00 6b 61 66 6b 61 2f 75 62 75 6e 74 75 32 36 2e 6d 73 74 6f 72 6d 2e 63 6f 6d 40 4d 53 54 4f 52 4d 2e 43 4f 4d fc 8c d0 79 4b e0 8a f1 ff 2a 40 c5 ] [2018-09-06 08:00:51,048] INFO zookeeper state changed (SaslAuthenticated) (org.I0Itec.zkclient.ZkClient) [2018-09-06 08:00:51,296] INFO Cluster ID = 9A-L_6nVSf2GswmsFHv9TA (kafka.server.KafkaServer) [2018-09-06 08:00:51,314] INFO [ThrottledRequestReaper-Fetch], Starting (kafka.server.ClientQuotaManager$ThrottledRequestReaper) [2018-09-06 08:00:51,315] INFO [ThrottledRequestReaper-Produce], Starting (kafka.server.ClientQuotaManager$ThrottledRequestReaper) [2018-09-06 08:00:51,340] INFO Loading logs. (kafka.log.LogManager) [2018-09-06 08:00:51,345] INFO Logs loading complete in 5 ms. (kafka.log.LogManager) [2018-09-06 08:00:51,397] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) [2018-09-06 08:00:51,400] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) Debug is true storeKey true useTicketCache false useKeyTab true doNotPrompt false ticketCache is null isInitiator true KeyTab is /tmp/kafka.keytab refreshKrb5Config is true principal is kafka/kafka1.hostname.com@MSTORM.COM tryFirstPass is false useFirstPass is false storePass is false clearPass is false Refreshing Kerberos configuration Java config name: /etc/krb5.conf Loaded from Java config >>> KdcAccessibility: reset Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. >>> KrbAsReq creating message >>> KrbKdcReq send: kdc=kerberosServer.hostname.com UDP:88, timeout=30000, number of retries =3, #bytes=159 >>> KDCCommunication: kdc=kerberosServer.hostname.com UDP:88, timeout=30000,Attempt =1, #bytes=159 >>> KrbKdcReq send: #bytes read=387 >>>Pre-Authentication Data: PA-DATA type = 136 >>>Pre-Authentication Data: PA-DATA type = 19 PA-ETYPE-INFO2 etype = 18, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 23, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 16, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null >>>Pre-Authentication Data: PA-DATA type = 2 PA-ENC-TIMESTAMP >>>Pre-Authentication Data: PA-DATA type = 133 >>> KdcAccessibility: remove kerberosServer.hostname.com >>> KDCRep: init() encoding tag is 126 req type is 11 >>>KRBError: cTime is Tue Apr 07 23:54:33 EDT 2015 1428465273000 sTime is Thu Sep 06 08:00:51 EDT 2018 1536235251000 suSec is 428427 error code is 25 error Message is Additional pre-authentication required cname is kafka/kafka1.hostname.com@MSTORM.COM sname is krbtgt/MSTORM.COM@MSTORM.COM eData provided. msgType is 30 >>>Pre-Authentication Data: PA-DATA type = 136 >>>Pre-Authentication Data: PA-DATA type = 19 PA-ETYPE-INFO2 etype = 18, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 23, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null PA-ETYPE-INFO2 etype = 16, salt = MSTORM.COMkafkakafka1.hostname.com, s2kparams = null >>>Pre-Authentication Data: PA-DATA type = 2 PA-ENC-TIMESTAMP >>>Pre-Authentication Data: PA-DATA type = 133 KRBError received: NEEDED_PREAUTH KrbAsReqBuilder: PREAUTH FAILED/REQ, re-send AS-REQ Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 Using builtin default etypes for default_tkt_enctypes default etypes for default_tkt_enctypes: 18 17 16 23. >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbAsReq creating message >>> KrbKdcReq send: kdc=kerberosServer.hostname.com UDP:88, timeout=30000, number of retries =3, #bytes=246 >>> KDCCommunication: kdc=kerberosServer.hostname.com UDP:88, timeout=30000,Attempt =1, #bytes=246 >>> KrbKdcReq send: #bytes read=707 >>> KdcAccessibility: remove kerberosServer.hostname.com Looking for keys for: kafka/kafka1.hostname.com@MSTORM.COM Found unsupported keytype (1) for kafka/kafka1.hostname.com@MSTORM.COM Added key: 16version: 2 Added key: 23version: 2 Added key: 18version: 2 >>> EType: sun.security.krb5.internal.crypto.Aes256CtsHmacSha1EType >>> KrbAsRep cons in KrbAsReq.getReply kafka/kafka1.hostname.com principal is kafka/kafka1.hostname.com@MSTORM.COM Will use keytab Commit Succeeded [2018-09-06 08:00:51,465] INFO Successfully logged in. (org.apache.kafka.common.security.authenticator.AbstractLogin) [2018-09-06 08:00:51,467] INFO [Principal=kafka/kafka1.hostname.com@MSTORM.COM]: TGT refresh thread started. (org.apache.kafka.common.security.kerberos.KerberosLogin) [2018-09-06 08:00:51,467] INFO [Principal=kafka/kafka1.hostname.com@MSTORM.COM]: TGT valid starting at: Thu Sep 06 08:00:51 EDT 2018 (org.apache.kafka.common.security.kerberos.KerberosLogin) [2018-09-06 08:00:51,467] INFO [Principal=kafka/kafka1.hostname.com@MSTORM.COM]: TGT expires: Thu Sep 06 18:00:51 EDT 2018 (org.apache.kafka.common.security.kerberos.KerberosLogin) [2018-09-06 08:00:51,467] INFO [Principal=kafka/kafka1.hostname.com@MSTORM.COM]: TGT refresh sleeping until: Thu Sep 06 16:09:45 EDT 2018 (org.apache.kafka.common.security.kerberos.KerberosLogin) [2018-09-06 08:00:51,499] INFO Awaiting socket connections on kafka1.hostname.com:9092. (kafka.network.Acceptor) [2018-09-06 08:00:51,501] INFO [Socket Server on Broker 1], Started 1 acceptor threads (kafka.network.SocketServer) [2018-09-06 08:00:51,517] INFO [ExpirationReaper-1], Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2018-09-06 08:00:51,518] INFO [ExpirationReaper-1], Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2018-09-06 08:00:51,543] INFO Creating /controller (is it secure? true) (kafka.utils.ZKCheckedEphemeral) [2018-09-06 08:00:51,555] ERROR Invalid ACL (kafka.utils.ZKCheckedEphemeral) [2018-09-06 08:00:51,556] INFO Result of znode creation is: INVALIDACL (kafka.utils.ZKCheckedEphemeral) [2018-09-06 08:00:51,557] ERROR Error while electing or becoming leader on broker 1 (kafka.server.ZookeeperLeaderElector) org.I0Itec.zkclient.exception.ZkException: org.apache.zookeeper.KeeperException$InvalidACLException: KeeperErrorCode = InvalidACL at org.I0Itec.zkclient.exception.ZkException.create(ZkException.java:68) at kafka.utils.ZKCheckedEphemeral.create(ZkUtils.scala:1124) at kafka.server.ZookeeperLeaderElector.elect(ZookeeperLeaderElector.scala:82) at kafka.server.ZookeeperLeaderElector.$anonfun$startup$1(ZookeeperLeaderElector.scala:51) at scala.runtime.java8.JFunction0$mcZ$sp.apply(JFunction0$mcZ$sp.java:12) at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:213) at kafka.server.ZookeeperLeaderElector.startup(ZookeeperLeaderElector.scala:49) at kafka.controller.KafkaController.$anonfun$startup$1(KafkaController.scala:681) at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:12) at kafka.utils.CoreUtils$.inLock(CoreUtils.scala:213) at kafka.controller.KafkaController.startup(KafkaController.scala:677) at kafka.server.KafkaServer.startup(KafkaServer.scala:224) at kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:39) at kafka.Kafka$.main(Kafka.scala:67) at kafka.Kafka.main(Kafka.scala) Caused by: org.apache.zookeeper.KeeperException$InvalidACLException: KeeperErrorCode = InvalidACL at org.apache.zookeeper.KeeperException.create(KeeperException.java:121) ... 14 more [2018-09-06 08:00:51,574] INFO [ExpirationReaper-1], Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2018-09-06 08:00:51,576] INFO [ExpirationReaper-1], Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2018-09-06 08:00:51,577] INFO [ExpirationReaper-1], Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) [2018-09-06 08:00:51,586] INFO [GroupCoordinator 1]: Starting up. (kafka.coordinator.GroupCoordinator) [2018-09-06 08:00:51,587] INFO [GroupCoordinator 1]: Startup complete. (kafka.coordinator.GroupCoordinator) [2018-09-06 08:00:51,588] INFO [Group Metadata Manager on Broker 1]: Removed 0 expired offsets in 2 milliseconds. (kafka.coordinator.GroupMetadataManager) [2018-09-06 08:00:51,601] INFO Will not load MX4J, mx4j-tools.jar is not in the classpath (kafka.utils.Mx4jLoader$) [2018-09-06 08:00:51,619] INFO Creating /brokers/ids/1 (is it secure? false) (kafka.utils.ZKCheckedEphemeral) [2018-09-06 08:00:51,638] INFO Result of znode creation is: OK (kafka.utils.ZKCheckedEphemeral) [2018-09-06 08:00:51,639] INFO Registered broker 1 at path /brokers/ids/1 with addresses: EndPoint(kafka1.hostname.com,9092,ListenerName(SASL_PLAINTEXT),SASL_PLAINTEXT) (kafka.utils.ZkUtils) [2018-09-06 08:00:51,652] INFO Kafka version : 0.10.2.0 (org.apache.kafka.common.utils.AppInfoParser) [2018-09-06 08:00:51,652] INFO Kafka commitId : 576d93a8dc0cf421 (org.apache.kafka.common.utils.AppInfoParser) [2018-09-06 08:00:51,653] INFO [Kafka Server 1], started (kafka.server.KafkaServer) [2018-09-06 08:10:51,586] INFO [Group Metadata Manager on Broker 1]: Removed 0 expired offsets in 0 milliseconds. (kafka.coordinator.GroupMetadataManager) [2018-09-06 08:20:51,586] INFO [Group Metadata Manager on Broker 1]: Removed 0 expired offsets in 0 milliseconds. (kafka.coordinator.GroupMetadataManager) [2018-09-06 08:30:51,586] INFO [Group Metadata Manager on Broker 1]: Removed 0 expired offsets in 0 milliseconds. (kafka.coordinator.GroupMetadataManager