2018-05-25 13:35:33.999 o.a.s.d.worker main [INFO] Launching worker for checkpoint_wifi-20-1527226482 on c2953577-5859-4402-96e9-1220aafde03a:6700 with id fb8b6655-e8fd-4916-a12c-f8ca6d724fd0 and conf {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "distributed", "topology.metrics.metric.name.separator" ".", "storm.messaging.netty.client_worker_threads" 1, "client.jartransformer.class" "org.apache.storm.hack.StormShadeTransformer", "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 3, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 5000, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2181, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "task.backpressure.poll.secs" 30, "drpc.queue.size" 128, "topology.backpressure.enable" false, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "nimbus.reassign" true, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["master1.nic.in" "master2.nic.in"], "topology.metrics.expand.map.type" true, "nimbus.thrift.threads" 196, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" "/etc/hbase/conf:/etc/hadoop/conf", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" false, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/hadoop/storm", "worker.childopts" "-Xmx768m", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m ", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.disable.symlinks" false, "storm.messaging.netty.max_retries" 30, "ui.childopts" "-Xmx768m ", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 30000, "drpc.childopts" "-Xmx768m ", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 30000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 1000, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "topology.metrics.aggregate.per.worker" true, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "storm.cluster.metrics.consumer.publish.interval.secs" 60, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" [6700 6701], "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "topology.optimize" true, "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "topology.metrics.aggregate.metric.evict.secs" 5, "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "topology.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink", "parallelism.hint" 1, "whitelist" ["kafkaOffset\\..+/" "__complete-latency" "__process-latency" "__receive\\.population$" "__sendqueue\\.population$" "__execute-count" "__emit-count" "__ack-count" "__fail-count" "memory/heap\\.usedBytes$" "memory/nonHeap\\.usedBytes$" "GC/.+\\.count$" "GC/.+\\.timeMs$"]}], "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "topology.disable.loadaware.messaging" false, "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["master1.nic.in"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "storm.blobstore.acl.validation.enabled" false, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.cluster.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsReporter"}], "storm.thrift.socket.timeout.ms" 600000, "supervisor.childopts" "-Xmx1024m", "topology.worker.max.heap.size.mb" 768.0, "ui.http.x-frame-options" "DENY", "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "topology.receiver.buffer.size" 8, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 500, "topology.disruptor.wait.strategy" "com.lmax.disruptor.BlockingWaitStrategy", "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib", "ui.port" 8744, "storm.log.dir" "/var/log/storm", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} 2018-05-25 13:35:34.175 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl main [INFO] Starting 2018-05-25 13:35:34.184 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:zookeeper.version=3.4.6-235--1, built on 10/30/2017 01:54 GMT 2018-05-25 13:35:34.185 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:host.name=metron1.nic.in 2018-05-25 13:35:34.185 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.version=1.8.0_112 2018-05-25 13:35:34.185 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.vendor=Oracle Corporation 2018-05-25 13:35:34.185 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.home=/usr/jdk64/jdk1.8.0_112/jre 2018-05-25 13:35:34.185 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.class.path=/usr/hdp/2.6.3.0-235/storm/lib/asm-5.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-api-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-core-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-slf4j-impl-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/objenesis-2.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/slf4j-api-1.7.21.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-core-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-rename-hack-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/zookeeper.jar:/usr/hdp/2.6.3.0-235/storm/lib/ambari-metrics-storm-sink.jar:/usr/hdp/current/storm-supervisor/conf:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/stormjar.jar:/etc/hbase/conf:/etc/hadoop/conf 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.library.path=/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources/Linux-amd64:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources:/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.io.tmpdir=/hadoop/storm/workers/fb8b6655-e8fd-4916-a12c-f8ca6d724fd0/tmp 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.compiler= 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.name=Linux 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.arch=amd64 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.version=3.10.0-693.17.1.el7.x86_64 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.name=storm 2018-05-25 13:35:34.186 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.home=/home/storm 2018-05-25 13:35:34.187 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.dir=/hadoop/storm/workers/fb8b6655-e8fd-4916-a12c-f8ca6d724fd0 2018-05-25 13:35:34.188 o.a.s.s.o.a.z.ZooKeeper main [INFO] Initiating client connection, connectString=master1.nic.in:2181,master2.nic.in:2181 sessionTimeout=30000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@c269425 2018-05-25 13:35:34.216 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Opening socket connection to server master2.nic.in/10.122.62.99:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:35:34.309 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:43306, server: master2.nic.in/10.122.62.99:2181 2018-05-25 13:35:34.327 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Session establishment complete on server master2.nic.in/10.122.62.99:2181, sessionid = 0x2638164225b164d, negotiated timeout = 30000 2018-05-25 13:35:34.335 o.a.s.s.o.a.c.f.s.ConnectionStateManager main-EventThread [INFO] State change: CONNECTED 2018-05-25 13:35:34.339 o.a.s.zookeeper main-EventThread [INFO] Zookeeper state update: :connected:none 2018-05-25 13:35:34.354 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl Curator-Framework-0 [INFO] backgroundOperationsLoop exiting 2018-05-25 13:35:34.361 o.a.s.s.o.a.z.ClientCnxn main-EventThread [INFO] EventThread shut down 2018-05-25 13:35:34.361 o.a.s.s.o.a.z.ZooKeeper main [INFO] Session: 0x2638164225b164d closed 2018-05-25 13:35:34.364 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl main [INFO] Starting 2018-05-25 13:35:34.365 o.a.s.s.o.a.z.ZooKeeper main [INFO] Initiating client connection, connectString=master1.nic.in:2181,master2.nic.in:2181/storm sessionTimeout=30000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5f95f1e1 2018-05-25 13:35:34.368 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Opening socket connection to server master1.nic.in/10.122.62.98:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:35:34.369 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:45600, server: master1.nic.in/10.122.62.98:2181 2018-05-25 13:35:34.373 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Session establishment complete on server master1.nic.in/10.122.62.98:2181, sessionid = 0x163816421312c2c, negotiated timeout = 30000 2018-05-25 13:35:34.373 o.a.s.s.o.a.c.f.s.ConnectionStateManager main-EventThread [INFO] State change: CONNECTED 2018-05-25 13:35:34.427 o.a.s.s.a.AuthUtils main [INFO] Got AutoCreds [] 2018-05-25 13:35:34.432 o.a.s.d.worker main [INFO] Reading Assignments. 2018-05-25 13:35:34.534 o.a.s.m.TransportFactory main [INFO] Storm peer transport plugin:org.apache.storm.messaging.netty.Context 2018-05-25 13:35:34.935 o.a.s.m.n.Server main [INFO] Create Netty Server Netty-server-localhost-6700, buffer_size: 5242880, maxWorkers: 1 2018-05-25 13:35:35.161 o.a.s.d.worker main [INFO] Registering IConnectionCallbacks for c2953577-5859-4402-96e9-1220aafde03a:6700 2018-05-25 13:35:35.230 o.a.s.d.executor main [INFO] Loading executor __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:35:35.253 o.a.s.d.executor main [INFO] Loaded executor tasks __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:35:35.269 o.a.s.d.executor main [INFO] Finished loading executor __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:35:35.287 o.a.s.d.executor main [INFO] Loading executor errorMessageWriter:[3 3] 2018-05-25 13:35:35.400 o.a.s.d.executor main [INFO] Loaded executor tasks errorMessageWriter:[3 3] 2018-05-25 13:35:35.404 o.a.s.d.executor main [INFO] Finished loading executor errorMessageWriter:[3 3] 2018-05-25 13:35:35.417 o.a.s.d.executor main [INFO] Loading executor __acker:[1 1] 2018-05-25 13:35:35.419 o.a.s.d.executor main [INFO] Loaded executor tasks __acker:[1 1] 2018-05-25 13:35:35.422 o.a.s.d.executor main [INFO] Timeouts disabled for executor __acker:[1 1] 2018-05-25 13:35:35.422 o.a.s.d.executor main [INFO] Finished loading executor __acker:[1 1] 2018-05-25 13:35:35.434 o.a.s.d.executor main [INFO] Loading executor __system:[-1 -1] 2018-05-25 13:35:35.436 o.a.s.d.executor main [INFO] Loaded executor tasks __system:[-1 -1] 2018-05-25 13:35:35.438 o.a.s.d.executor main [INFO] Finished loading executor __system:[-1 -1] 2018-05-25 13:35:35.452 o.a.s.d.executor main [INFO] Loading executor parserBolt:[5 5] 2018-05-25 13:35:35.480 o.a.s.d.executor main [INFO] Loaded executor tasks parserBolt:[5 5] 2018-05-25 13:35:35.483 o.a.s.d.executor main [INFO] Finished loading executor parserBolt:[5 5] 2018-05-25 13:35:35.494 o.a.s.d.executor main [INFO] Loading executor kafkaSpout:[4 4] 2018-05-25 13:35:35.550 o.a.s.d.executor main [INFO] Loaded executor tasks kafkaSpout:[4 4] 2018-05-25 13:35:35.564 o.a.s.d.executor main [INFO] Finished loading executor kafkaSpout:[4 4] 2018-05-25 13:35:35.581 o.a.s.d.worker main [INFO] Started with log levels: {"" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "STDERR" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "STDOUT" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "org.apache.storm.metric.LoggingMetricsConsumer" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"]} 2018-05-25 13:35:35.596 o.a.s.d.worker main [INFO] Worker has topology config {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "distributed", "topology.metrics.metric.name.separator" ".", "storm.messaging.netty.client_worker_threads" 1, "client.jartransformer.class" "org.apache.storm.hack.StormShadeTransformer", "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 3, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 5000, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2181, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "task.backpressure.poll.secs" 30, "drpc.queue.size" 128, "topology.backpressure.enable" false, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "nimbus.reassign" true, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "topology.submitter.principal" "", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["master1.nic.in" "master2.nic.in"], "topology.metrics.expand.map.type" true, "nimbus.thrift.threads" 196, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" "/etc/hbase/conf:/etc/hadoop/conf", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" false, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/hadoop/storm", "worker.childopts" "-Xmx768m", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m ", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.disable.symlinks" false, "storm.messaging.netty.max_retries" 30, "ui.childopts" "-Xmx768m ", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 30000, "drpc.childopts" "-Xmx768m ", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 30000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 1000, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "storm.zookeeper.superACL" nil, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "topology.metrics.aggregate.per.worker" true, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "storm.cluster.metrics.consumer.publish.interval.secs" 60, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" [6700 6701], "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "topology.optimize" true, "topology.users" [], "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "topology.metrics.aggregate.metric.evict.secs" 5, "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.submitter.user" "storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "topology.metrics.consumer.register" [{"whitelist" ["kafkaOffset\\..+/" "__complete-latency" "__process-latency" "__receive\\.population$" "__sendqueue\\.population$" "__execute-count" "__emit-count" "__ack-count" "__fail-count" "memory/heap\\.usedBytes$" "memory/nonHeap\\.usedBytes$" "GC/.+\\.count$" "GC/.+\\.timeMs$"], "class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink", "parallelism.hint" 1}], "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "topology.disable.loadaware.messaging" false, "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "topology.kryo.register" nil, "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["master1.nic.in"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "storm.blobstore.acl.validation.enabled" false, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.cluster.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsReporter"}], "storm.thrift.socket.timeout.ms" 600000, "supervisor.childopts" "-Xmx1024m", "topology.worker.max.heap.size.mb" 768.0, "ui.http.x-frame-options" "DENY", "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "topology.receiver.buffer.size" 8, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 500, "topology.disruptor.wait.strategy" "com.lmax.disruptor.BlockingWaitStrategy", "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib", "ui.port" 8744, "storm.log.dir" "/var/log/storm", "topology.kryo.decorators" [], "storm.id" "checkpoint_wifi-20-1527226482", "topology.name" "checkpoint_wifi", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} 2018-05-25 13:35:35.597 o.a.s.d.worker main [INFO] Worker fb8b6655-e8fd-4916-a12c-f8ca6d724fd0 for storm checkpoint_wifi-20-1527226482 on c2953577-5859-4402-96e9-1220aafde03a:6700 has finished loading 2018-05-25 13:35:36.109 o.a.s.d.worker refresh-active-timer [INFO] All connections are ready for worker c2953577-5859-4402-96e9-1220aafde03a:6700 with id fb8b6655-e8fd-4916-a12c-f8ca6d724fd0 2018-05-25 13:35:36.143 o.a.s.d.executor Thread-10-__system-executor[-1 -1] [INFO] Preparing bolt __system:(-1) 2018-05-25 13:35:36.158 o.a.s.d.executor Thread-10-__system-executor[-1 -1] [INFO] Prepared bolt __system:(-1) 2018-05-25 13:35:36.165 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Opening spout kafkaSpout:(4) 2018-05-25 13:35:36.170 o.a.s.d.executor Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Preparing bolt __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:(2) 2018-05-25 13:35:36.171 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka Spout opened with the following configuration: KafkaSpoutConfig{kafkaProps={security.protocol=PLAINTEXT, enable.auto.commit=false, group.id=checkpoint_wifi_parser, bootstrap.servers=master2.nic.in:6667,master1.nic.in:6667}, key=org.apache.kafka.common.serialization.ByteArrayDeserializer@3bf4923c, value=org.apache.kafka.common.serialization.ByteArrayDeserializer@42598547, pollTimeoutMs=200, offsetCommitPeriodMs=30000, maxUncommittedOffsets=10000000, firstPollOffsetStrategy=UNCOMMITTED_EARLIEST, subscription=org.apache.storm.kafka.spout.NamedSubscription@6a9ba48c, translator=org.apache.metron.storm.kafka.flux.SimpleStormKafkaBuilder$SpoutRecordTranslator@48da909e, retryService=KafkaSpoutRetryExponentialBackoff{delay=TimeInterval{length=0, timeUnit=SECONDS}, ratio=TimeInterval{length=2, timeUnit=MILLISECONDS}, maxRetries=2147483647, maxRetryDelay=TimeInterval{length=10, timeUnit=SECONDS}}} 2018-05-25 13:35:36.173 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Opened spout kafkaSpout:(4) 2018-05-25 13:35:36.177 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Activating spout kafkaSpout:(4) 2018-05-25 13:35:36.185 o.a.s.d.executor Thread-12-parserBolt-executor[5 5] [INFO] Preparing bolt parserBolt:(5) 2018-05-25 13:35:36.205 o.a.s.d.executor Thread-6-errorMessageWriter-executor[3 3] [INFO] Preparing bolt errorMessageWriter:(3) 2018-05-25 13:35:36.218 o.a.k.c.c.ConsumerConfig Thread-14-kafkaSpout-executor[4 4] [INFO] ConsumerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor] reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 max.partition.fetch.bytes = 1048576 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS enable.auto.commit = false sasl.mechanism = GSSAPI interceptor.classes = null exclude.internal.topics = true ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null max.poll.records = 2147483647 check.crcs = true request.timeout.ms = 40000 heartbeat.interval.ms = 3000 auto.commit.interval.ms = 5000 receive.buffer.bytes = 65536 ssl.truststore.type = JKS ssl.truststore.location = null ssl.keystore.password = null fetch.min.bytes = 1 send.buffer.bytes = 131072 value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer group.id = checkpoint_wifi_parser retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX ssl.key.password = null fetch.max.wait.ms = 500 sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 session.timeout.ms = 30000 metrics.num.samples = 2 key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 auto.offset.reset = latest 2018-05-25 13:35:36.226 o.a.s.d.executor Thread-8-__acker-executor[1 1] [INFO] Preparing bolt __acker:(1) 2018-05-25 13:35:36.232 o.a.s.d.executor Thread-8-__acker-executor[1 1] [INFO] Prepared bolt __acker:(1) 2018-05-25 13:35:36.243 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:35:36.290 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = producer-1 ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:35:36.290 o.a.k.c.c.ConsumerConfig Thread-14-kafkaSpout-executor[4 4] [INFO] ConsumerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor] reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 max.partition.fetch.bytes = 1048576 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS enable.auto.commit = false sasl.mechanism = GSSAPI interceptor.classes = null exclude.internal.topics = true ssl.truststore.password = null client.id = consumer-1 ssl.endpoint.identification.algorithm = null max.poll.records = 2147483647 check.crcs = true request.timeout.ms = 40000 heartbeat.interval.ms = 3000 auto.commit.interval.ms = 5000 receive.buffer.bytes = 65536 ssl.truststore.type = JKS ssl.truststore.location = null ssl.keystore.password = null fetch.min.bytes = 1 send.buffer.bytes = 131072 value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer group.id = checkpoint_wifi_parser retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX ssl.key.password = null fetch.max.wait.ms = 500 sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 session.timeout.ms = 30000 metrics.num.samples = 2 key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 auto.offset.reset = latest 2018-05-25 13:35:36.292 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [WARN] The configuration request.required.acks = 1 was supplied but isn't a known config. 2018-05-25 13:35:36.296 o.a.k.c.u.AppInfoParser Thread-6-errorMessageWriter-executor[3 3] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:35:36.296 o.a.k.c.u.AppInfoParser Thread-6-errorMessageWriter-executor[3 3] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:35:36.299 o.a.s.d.executor Thread-6-errorMessageWriter-executor[3 3] [INFO] Prepared bolt errorMessageWriter:(3) 2018-05-25 13:35:36.342 o.a.k.c.u.AppInfoParser Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:35:36.343 o.a.k.c.u.AppInfoParser Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:35:36.345 o.a.s.k.s.NamedSubscription Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka consumer subscribed topics [checkpoint_wifi] 2018-05-25 13:35:36.363 o.a.c.f.i.CuratorFrameworkImpl Thread-12-parserBolt-executor[5 5] [INFO] Starting 2018-05-25 13:35:36.378 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:zookeeper.version=3.4.6-235--1, built on 10/30/2017 02:17 GMT 2018-05-25 13:35:36.378 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:host.name=metron1.nic.in 2018-05-25 13:35:36.378 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.version=1.8.0_112 2018-05-25 13:35:36.378 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.vendor=Oracle Corporation 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.home=/usr/jdk64/jdk1.8.0_112/jre 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.class.path=/usr/hdp/2.6.3.0-235/storm/lib/asm-5.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-api-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-core-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-slf4j-impl-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/objenesis-2.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/slf4j-api-1.7.21.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-core-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-rename-hack-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/zookeeper.jar:/usr/hdp/2.6.3.0-235/storm/lib/ambari-metrics-storm-sink.jar:/usr/hdp/current/storm-supervisor/conf:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/stormjar.jar:/etc/hbase/conf:/etc/hadoop/conf 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.library.path=/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources/Linux-amd64:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources:/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.io.tmpdir=/hadoop/storm/workers/fb8b6655-e8fd-4916-a12c-f8ca6d724fd0/tmp 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.compiler= 2018-05-25 13:35:36.379 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.name=Linux 2018-05-25 13:35:36.380 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.arch=amd64 2018-05-25 13:35:36.380 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.version=3.10.0-693.17.1.el7.x86_64 2018-05-25 13:35:36.380 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.name=storm 2018-05-25 13:35:36.380 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.home=/home/storm 2018-05-25 13:35:36.380 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.dir=/hadoop/storm/workers/fb8b6655-e8fd-4916-a12c-f8ca6d724fd0 2018-05-25 13:35:36.381 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Initiating client connection, connectString=master2.nic.in:2181,master1.nic.in:2181 sessionTimeout=60000 watcher=org.apache.curator.ConnectionState@267ce37 2018-05-25 13:35:36.405 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Opening socket connection to server master1.nic.in/10.122.62.98:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:35:36.407 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:45612, server: master1.nic.in/10.122.62.98:2181 2018-05-25 13:35:36.417 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Session establishment complete on server master1.nic.in/10.122.62.98:2181, sessionid = 0x163816421312c2d, negotiated timeout = 60000 2018-05-25 13:35:36.431 o.a.c.f.s.ConnectionStateManager Thread-12-parserBolt-executor[5 5]-EventThread [INFO] State change: CONNECTED 2018-05-25 13:35:36.464 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Discovered coordinator master1.nic.in:6667 (id: 2147482646 rack: null) for group checkpoint_wifi_parser. 2018-05-25 13:35:36.470 o.a.k.c.c.i.ConsumerCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Revoking previously assigned partitions [] for group checkpoint_wifi_parser 2018-05-25 13:35:36.472 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Partitions revoked. [consumer-group=checkpoint_wifi_parser, consumer=org.apache.kafka.clients.consumer.KafkaConsumer@527ee310, topic-partitions=[]] 2018-05-25 13:35:36.472 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] (Re-)joining group checkpoint_wifi_parser 2018-05-25 13:35:36.477 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Preparing Storm Metrics Sink 2018-05-25 13:35:36.491 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Successfully joined group checkpoint_wifi_parser with generation 9 2018-05-25 13:35:36.493 o.a.k.c.c.i.ConsumerCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Setting newly assigned partitions [checkpoint_wifi-0] for group checkpoint_wifi_parser 2018-05-25 13:35:36.493 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Partitions reassignment. [consumer-group=checkpoint_wifi_parser, consumer=org.apache.kafka.clients.consumer.KafkaConsumer@527ee310, topic-partitions=[checkpoint_wifi-0]] 2018-05-25 13:35:36.494 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Topology name "checkpoint_wifi" contains "_" which can be problematic for AMS. 2018-05-25 13:35:36.494 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Encouraged to not using any of these strings: ".", "_" 2018-05-25 13:35:36.494 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Same suggestion applies to component name. 2018-05-25 13:35:36.496 o.a.s.d.executor Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Prepared bolt __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:(2) 2018-05-25 13:35:36.508 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Initialization complete 2018-05-25 13:35:37.123 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:35:37.128 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = producer-2 ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:35:37.128 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [WARN] The configuration request.required.acks = 1 was supplied but isn't a known config. 2018-05-25 13:35:37.129 o.a.k.c.u.AppInfoParser Thread-12-parserBolt-executor[5 5] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:35:37.129 o.a.k.c.u.AppInfoParser Thread-12-parserBolt-executor[5 5] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:35:37.130 o.a.s.d.executor Thread-12-parserBolt-executor[5 5] [INFO] Prepared bolt parserBolt:(5) 2018-05-25 13:37:37.184 o.a.h.m.s.t.a.MetricSinkWriteShardHostnameHashingStrategy Thread-19 [INFO] Calculated collector shard master2.nic.in based on hostname: metron1.nic.in 2018-05-25 13:49:11.705 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Marking the coordinator master1.nic.in:6667 (id: 2147482646 rack: null) dead for group checkpoint_wifi_parser 2018-05-25 13:49:13.021 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Discovered coordinator master1.nic.in:6667 (id: 2147482646 rack: null) for group checkpoint_wifi_parser. 2018-05-25 13:49:13.062 o.a.k.c.c.i.ConsumerCoordinator Thread-14-kafkaSpout-executor[4 4] [ERROR] Group checkpoint_wifi_parser failed to commit partition checkpoint_wifi-0 at offset 3283224: This server does not host this topic-partition. 2018-05-25 13:49:13.065 o.a.s.util Thread-14-kafkaSpout-executor[4 4] [ERROR] Async loop died! org.apache.kafka.common.KafkaException: Unexpected error in commit: This server does not host this topic-partition. at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:609) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:541) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:679) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:658) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture$1.onSuccess(RequestFuture.java:167) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:133) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient$RequestFutureCompletionHandler.onComplete(ConsumerNetworkClient.java:426) ~[stormjar.jar:?] at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:278) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.clientPoll(ConsumerNetworkClient.java:360) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:224) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:192) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:163) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.commitOffsetsSync(ConsumerCoordinator.java:426) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.KafkaConsumer.commitSync(KafkaConsumer.java:1059) ~[stormjar.jar:?] at org.apache.storm.kafka.spout.KafkaSpout.commitOffsetsForAckedTuples(KafkaSpout.java:384) ~[stormjar.jar:?] at org.apache.storm.kafka.spout.KafkaSpout.nextTuple(KafkaSpout.java:219) ~[stormjar.jar:?] at org.apache.storm.daemon.executor$fn__10125$fn__10140$fn__10171.invoke(executor.clj:647) ~[storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at org.apache.storm.util$async_loop$fn__1221.invoke(util.clj:484) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_112] 2018-05-25 13:49:13.072 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [ERROR] org.apache.kafka.common.KafkaException: Unexpected error in commit: This server does not host this topic-partition. at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:609) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator$OffsetCommitResponseHandler.handle(ConsumerCoordinator.java:541) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:679) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.AbstractCoordinator$CoordinatorResponseHandler.onSuccess(AbstractCoordinator.java:658) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture$1.onSuccess(RequestFuture.java:167) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture.fireSuccess(RequestFuture.java:133) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.RequestFuture.complete(RequestFuture.java:107) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient$RequestFutureCompletionHandler.onComplete(ConsumerNetworkClient.java:426) ~[stormjar.jar:?] at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:278) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.clientPoll(ConsumerNetworkClient.java:360) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:224) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:192) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:163) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.internals.ConsumerCoordinator.commitOffsetsSync(ConsumerCoordinator.java:426) ~[stormjar.jar:?] at org.apache.kafka.clients.consumer.KafkaConsumer.commitSync(KafkaConsumer.java:1059) ~[stormjar.jar:?] at org.apache.storm.kafka.spout.KafkaSpout.commitOffsetsForAckedTuples(KafkaSpout.java:384) ~[stormjar.jar:?] at org.apache.storm.kafka.spout.KafkaSpout.nextTuple(KafkaSpout.java:219) ~[stormjar.jar:?] at org.apache.storm.daemon.executor$fn__10125$fn__10140$fn__10171.invoke(executor.clj:647) ~[storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at org.apache.storm.util$async_loop$fn__1221.invoke(util.clj:484) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_112] 2018-05-25 13:49:13.101 o.a.s.util Thread-14-kafkaSpout-executor[4 4] [ERROR] Halting process: ("Worker died") java.lang.RuntimeException: ("Worker died") at org.apache.storm.util$exit_process_BANG_.doInvoke(util.clj:341) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at clojure.lang.RestFn.invoke(RestFn.java:423) [clojure-1.7.0.jar:?] at org.apache.storm.daemon.worker$fn__10795$fn__10796.invoke(worker.clj:763) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at org.apache.storm.daemon.executor$mk_executor_data$fn__10011$fn__10012.invoke(executor.clj:276) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at org.apache.storm.util$async_loop$fn__1221.invoke(util.clj:494) [storm-core-1.1.0.2.6.3.0-235.jar:1.1.0.2.6.3.0-235] at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_112] 2018-05-25 13:49:13.104 o.a.s.d.worker Thread-17 [INFO] Shutting down worker checkpoint_wifi-20-1527226482 c2953577-5859-4402-96e9-1220aafde03a 6700 2018-05-25 13:49:13.104 o.a.s.d.worker Thread-17 [INFO] Terminating messaging context 2018-05-25 13:49:13.105 o.a.s.d.worker Thread-17 [INFO] Shutting down executors 2018-05-25 13:49:13.105 o.a.s.d.executor Thread-17 [INFO] Shutting down executor __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:49:13.107 o.a.s.util Thread-3-disruptor-executor[2 2]-send-queue [INFO] Async loop interrupted! 2018-05-25 13:49:13.108 o.a.s.util Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Async loop interrupted! 2018-05-25 13:49:22.469 o.a.s.d.worker main [INFO] Launching worker for checkpoint_wifi-20-1527226482 on c2953577-5859-4402-96e9-1220aafde03a:6700 with id d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a and conf {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "distributed", "topology.metrics.metric.name.separator" ".", "storm.messaging.netty.client_worker_threads" 1, "client.jartransformer.class" "org.apache.storm.hack.StormShadeTransformer", "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 3, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 5000, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2181, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "task.backpressure.poll.secs" 30, "drpc.queue.size" 128, "topology.backpressure.enable" false, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "nimbus.reassign" true, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["master1.nic.in" "master2.nic.in"], "topology.metrics.expand.map.type" true, "nimbus.thrift.threads" 196, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" "/etc/hbase/conf:/etc/hadoop/conf", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" false, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/hadoop/storm", "worker.childopts" "-Xmx768m", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m ", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.disable.symlinks" false, "storm.messaging.netty.max_retries" 30, "ui.childopts" "-Xmx768m ", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 30000, "drpc.childopts" "-Xmx768m ", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 30000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 1000, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "topology.metrics.aggregate.per.worker" true, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "storm.cluster.metrics.consumer.publish.interval.secs" 60, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" [6700 6701], "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "topology.optimize" true, "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "topology.metrics.aggregate.metric.evict.secs" 5, "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "topology.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink", "parallelism.hint" 1, "whitelist" ["kafkaOffset\\..+/" "__complete-latency" "__process-latency" "__receive\\.population$" "__sendqueue\\.population$" "__execute-count" "__emit-count" "__ack-count" "__fail-count" "memory/heap\\.usedBytes$" "memory/nonHeap\\.usedBytes$" "GC/.+\\.count$" "GC/.+\\.timeMs$"]}], "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "topology.disable.loadaware.messaging" false, "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["master1.nic.in"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "storm.blobstore.acl.validation.enabled" false, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.cluster.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsReporter"}], "storm.thrift.socket.timeout.ms" 600000, "supervisor.childopts" "-Xmx1024m", "topology.worker.max.heap.size.mb" 768.0, "ui.http.x-frame-options" "DENY", "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "topology.receiver.buffer.size" 8, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 500, "topology.disruptor.wait.strategy" "com.lmax.disruptor.BlockingWaitStrategy", "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib", "ui.port" 8744, "storm.log.dir" "/var/log/storm", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} 2018-05-25 13:49:22.632 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl main [INFO] Starting 2018-05-25 13:49:22.641 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:zookeeper.version=3.4.6-235--1, built on 10/30/2017 01:54 GMT 2018-05-25 13:49:22.641 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:host.name=metron1.nic.in 2018-05-25 13:49:22.641 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.version=1.8.0_112 2018-05-25 13:49:22.641 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.vendor=Oracle Corporation 2018-05-25 13:49:22.641 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.home=/usr/jdk64/jdk1.8.0_112/jre 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.class.path=/usr/hdp/2.6.3.0-235/storm/lib/asm-5.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-api-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-core-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-slf4j-impl-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/objenesis-2.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/slf4j-api-1.7.21.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-core-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-rename-hack-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/zookeeper.jar:/usr/hdp/2.6.3.0-235/storm/lib/ambari-metrics-storm-sink.jar:/usr/hdp/current/storm-supervisor/conf:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/stormjar.jar:/etc/hbase/conf:/etc/hadoop/conf 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.library.path=/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources/Linux-amd64:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources:/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.io.tmpdir=/hadoop/storm/workers/d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a/tmp 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:java.compiler= 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.name=Linux 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.arch=amd64 2018-05-25 13:49:22.642 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:os.version=3.10.0-693.17.1.el7.x86_64 2018-05-25 13:49:22.643 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.name=storm 2018-05-25 13:49:22.643 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.home=/home/storm 2018-05-25 13:49:22.643 o.a.s.s.o.a.z.ZooKeeper main [INFO] Client environment:user.dir=/hadoop/storm/workers/d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a 2018-05-25 13:49:22.644 o.a.s.s.o.a.z.ZooKeeper main [INFO] Initiating client connection, connectString=master1.nic.in:2181,master2.nic.in:2181 sessionTimeout=30000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@c269425 2018-05-25 13:49:22.673 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Opening socket connection to server master2.nic.in/10.122.62.99:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:49:22.768 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:45688, server: master2.nic.in/10.122.62.99:2181 2018-05-25 13:49:22.781 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master2.nic.in:2181) [INFO] Session establishment complete on server master2.nic.in/10.122.62.99:2181, sessionid = 0x2638164225b164e, negotiated timeout = 30000 2018-05-25 13:49:22.796 o.a.s.s.o.a.c.f.s.ConnectionStateManager main-EventThread [INFO] State change: CONNECTED 2018-05-25 13:49:22.800 o.a.s.zookeeper main-EventThread [INFO] Zookeeper state update: :connected:none 2018-05-25 13:49:22.815 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl Curator-Framework-0 [INFO] backgroundOperationsLoop exiting 2018-05-25 13:49:22.823 o.a.s.s.o.a.z.ClientCnxn main-EventThread [INFO] EventThread shut down 2018-05-25 13:49:22.823 o.a.s.s.o.a.z.ZooKeeper main [INFO] Session: 0x2638164225b164e closed 2018-05-25 13:49:22.827 o.a.s.s.o.a.c.f.i.CuratorFrameworkImpl main [INFO] Starting 2018-05-25 13:49:22.828 o.a.s.s.o.a.z.ZooKeeper main [INFO] Initiating client connection, connectString=master1.nic.in:2181,master2.nic.in:2181/storm sessionTimeout=30000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5f95f1e1 2018-05-25 13:49:22.830 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Opening socket connection to server master1.nic.in/10.122.62.98:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:49:22.832 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:47982, server: master1.nic.in/10.122.62.98:2181 2018-05-25 13:49:22.835 o.a.s.s.o.a.z.ClientCnxn main-SendThread(master1.nic.in:2181) [INFO] Session establishment complete on server master1.nic.in/10.122.62.98:2181, sessionid = 0x163816421312c2f, negotiated timeout = 30000 2018-05-25 13:49:22.835 o.a.s.s.o.a.c.f.s.ConnectionStateManager main-EventThread [INFO] State change: CONNECTED 2018-05-25 13:49:22.891 o.a.s.s.a.AuthUtils main [INFO] Got AutoCreds [] 2018-05-25 13:49:22.896 o.a.s.d.worker main [INFO] Reading Assignments. 2018-05-25 13:49:22.990 o.a.s.m.TransportFactory main [INFO] Storm peer transport plugin:org.apache.storm.messaging.netty.Context 2018-05-25 13:49:23.385 o.a.s.m.n.Server main [INFO] Create Netty Server Netty-server-localhost-6700, buffer_size: 5242880, maxWorkers: 1 2018-05-25 13:49:23.599 o.a.s.d.worker main [INFO] Registering IConnectionCallbacks for c2953577-5859-4402-96e9-1220aafde03a:6700 2018-05-25 13:49:23.670 o.a.s.d.executor main [INFO] Loading executor __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:49:23.694 o.a.s.d.executor main [INFO] Loaded executor tasks __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:49:23.710 o.a.s.d.executor main [INFO] Finished loading executor __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:[2 2] 2018-05-25 13:49:23.729 o.a.s.d.executor main [INFO] Loading executor errorMessageWriter:[3 3] 2018-05-25 13:49:23.821 o.a.s.d.executor main [INFO] Loaded executor tasks errorMessageWriter:[3 3] 2018-05-25 13:49:23.825 o.a.s.d.executor main [INFO] Finished loading executor errorMessageWriter:[3 3] 2018-05-25 13:49:23.838 o.a.s.d.executor main [INFO] Loading executor __acker:[1 1] 2018-05-25 13:49:23.839 o.a.s.d.executor main [INFO] Loaded executor tasks __acker:[1 1] 2018-05-25 13:49:23.842 o.a.s.d.executor main [INFO] Timeouts disabled for executor __acker:[1 1] 2018-05-25 13:49:23.842 o.a.s.d.executor main [INFO] Finished loading executor __acker:[1 1] 2018-05-25 13:49:23.854 o.a.s.d.executor main [INFO] Loading executor __system:[-1 -1] 2018-05-25 13:49:23.856 o.a.s.d.executor main [INFO] Loaded executor tasks __system:[-1 -1] 2018-05-25 13:49:23.858 o.a.s.d.executor main [INFO] Finished loading executor __system:[-1 -1] 2018-05-25 13:49:23.872 o.a.s.d.executor main [INFO] Loading executor parserBolt:[5 5] 2018-05-25 13:49:23.902 o.a.s.d.executor main [INFO] Loaded executor tasks parserBolt:[5 5] 2018-05-25 13:49:23.906 o.a.s.d.executor main [INFO] Finished loading executor parserBolt:[5 5] 2018-05-25 13:49:23.918 o.a.s.d.executor main [INFO] Loading executor kafkaSpout:[4 4] 2018-05-25 13:49:23.978 o.a.s.d.executor main [INFO] Loaded executor tasks kafkaSpout:[4 4] 2018-05-25 13:49:23.991 o.a.s.d.executor main [INFO] Finished loading executor kafkaSpout:[4 4] 2018-05-25 13:49:24.009 o.a.s.d.worker main [INFO] Started with log levels: {"" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "STDERR" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "STDOUT" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"], "org.apache.storm.metric.LoggingMetricsConsumer" #object[org.apache.logging.log4j.Level 0x3c9ef6e9 "INFO"]} 2018-05-25 13:49:24.025 o.a.s.d.worker main [INFO] Worker has topology config {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "distributed", "topology.metrics.metric.name.separator" ".", "storm.messaging.netty.client_worker_threads" 1, "client.jartransformer.class" "org.apache.storm.hack.StormShadeTransformer", "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 3, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 5000, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2181, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" true, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "task.backpressure.poll.secs" 30, "drpc.queue.size" 128, "topology.backpressure.enable" false, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "nimbus.reassign" true, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "topology.submitter.principal" "", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" nil, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["master1.nic.in" "master2.nic.in"], "topology.metrics.expand.map.type" true, "nimbus.thrift.threads" 196, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" "/etc/hbase/conf:/etc/hadoop/conf", "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" false, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/hadoop/storm", "worker.childopts" "-Xmx768m", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m ", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.disable.symlinks" false, "storm.messaging.netty.max_retries" 30, "ui.childopts" "-Xmx768m ", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 30000, "drpc.childopts" "-Xmx768m ", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 30000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 1000, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "nimbus.task.timeout.secs" 30, "storm.zookeeper.superACL" nil, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "topology.metrics.aggregate.per.worker" true, "storm.zookeeper.retry.interval" 1000, "logs.users" nil, "storm.cluster.metrics.consumer.publish.interval.secs" 60, "worker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "supervisor.slots.ports" [6700 6701], "topology.transfer.buffer.size" 1024, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "topology.optimize" true, "topology.users" [], "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.messaging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "topology.metrics.aggregate.metric.evict.secs" 5, "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.submitter.user" "storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "topology.metrics.consumer.register" [{"whitelist" ["kafkaOffset\\..+/" "__complete-latency" "__process-latency" "__receive\\.population$" "__sendqueue\\.population$" "__execute-count" "__emit-count" "__ack-count" "__fail-count" "memory/heap\\.usedBytes$" "memory/nonHeap\\.usedBytes$" "GC/.+\\.count$" "GC/.+\\.timeMs$"], "class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink", "parallelism.hint" 1}], "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "topology.disable.loadaware.messaging" false, "storm.messaging.netty.authentication" false, "topology.component.resources.onheap.memory.mb" 128.0, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "topology.kryo.register" nil, "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["master1.nic.in"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 768, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "storm.blobstore.acl.validation.enabled" false, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.cluster.metrics.consumer.register" [{"class" "org.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsReporter"}], "storm.thrift.socket.timeout.ms" 600000, "supervisor.childopts" "-Xmx1024m", "topology.worker.max.heap.size.mb" 768.0, "ui.http.x-frame-options" "DENY", "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "topology.receiver.buffer.size" 8, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 500, "topology.disruptor.wait.strategy" "com.lmax.disruptor.BlockingWaitStrategy", "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib", "ui.port" 8744, "storm.log.dir" "/var/log/storm", "topology.kryo.decorators" [], "storm.id" "checkpoint_wifi-20-1527226482", "topology.name" "checkpoint_wifi", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} 2018-05-25 13:49:24.027 o.a.s.d.worker main [INFO] Worker d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a for storm checkpoint_wifi-20-1527226482 on c2953577-5859-4402-96e9-1220aafde03a:6700 has finished loading 2018-05-25 13:49:24.551 o.a.s.d.worker refresh-active-timer [INFO] All connections are ready for worker c2953577-5859-4402-96e9-1220aafde03a:6700 with id d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a 2018-05-25 13:49:24.597 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Opening spout kafkaSpout:(4) 2018-05-25 13:49:24.603 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka Spout opened with the following configuration: KafkaSpoutConfig{kafkaProps={security.protocol=PLAINTEXT, enable.auto.commit=false, group.id=checkpoint_wifi_parser, bootstrap.servers=master2.nic.in:6667,master1.nic.in:6667}, key=org.apache.kafka.common.serialization.ByteArrayDeserializer@2d87e875, value=org.apache.kafka.common.serialization.ByteArrayDeserializer@574ce4e, pollTimeoutMs=200, offsetCommitPeriodMs=30000, maxUncommittedOffsets=10000000, firstPollOffsetStrategy=UNCOMMITTED_EARLIEST, subscription=org.apache.storm.kafka.spout.NamedSubscription@46f6168a, translator=org.apache.metron.storm.kafka.flux.SimpleStormKafkaBuilder$SpoutRecordTranslator@3c08f9dd, retryService=KafkaSpoutRetryExponentialBackoff{delay=TimeInterval{length=0, timeUnit=SECONDS}, ratio=TimeInterval{length=2, timeUnit=MILLISECONDS}, maxRetries=2147483647, maxRetryDelay=TimeInterval{length=10, timeUnit=SECONDS}}} 2018-05-25 13:49:24.604 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Opened spout kafkaSpout:(4) 2018-05-25 13:49:24.608 o.a.s.d.executor Thread-12-parserBolt-executor[5 5] [INFO] Preparing bolt parserBolt:(5) 2018-05-25 13:49:24.609 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Activating spout kafkaSpout:(4) 2018-05-25 13:49:24.612 o.a.s.d.executor Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Preparing bolt __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:(2) 2018-05-25 13:49:24.627 o.a.s.d.executor Thread-6-errorMessageWriter-executor[3 3] [INFO] Preparing bolt errorMessageWriter:(3) 2018-05-25 13:49:24.646 o.a.s.d.executor Thread-8-__acker-executor[1 1] [INFO] Preparing bolt __acker:(1) 2018-05-25 13:49:24.650 o.a.s.d.executor Thread-8-__acker-executor[1 1] [INFO] Prepared bolt __acker:(1) 2018-05-25 13:49:24.654 o.a.k.c.c.ConsumerConfig Thread-14-kafkaSpout-executor[4 4] [INFO] ConsumerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor] reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 max.partition.fetch.bytes = 1048576 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS enable.auto.commit = false sasl.mechanism = GSSAPI interceptor.classes = null exclude.internal.topics = true ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null max.poll.records = 2147483647 check.crcs = true request.timeout.ms = 40000 heartbeat.interval.ms = 3000 auto.commit.interval.ms = 5000 receive.buffer.bytes = 65536 ssl.truststore.type = JKS ssl.truststore.location = null ssl.keystore.password = null fetch.min.bytes = 1 send.buffer.bytes = 131072 value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer group.id = checkpoint_wifi_parser retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX ssl.key.password = null fetch.max.wait.ms = 500 sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 session.timeout.ms = 30000 metrics.num.samples = 2 key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 auto.offset.reset = latest 2018-05-25 13:49:24.660 o.a.s.d.executor Thread-10-__system-executor[-1 -1] [INFO] Preparing bolt __system:(-1) 2018-05-25 13:49:24.665 o.a.s.d.executor Thread-10-__system-executor[-1 -1] [INFO] Prepared bolt __system:(-1) 2018-05-25 13:49:24.666 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:49:24.732 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = producer-1 ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:49:24.734 o.a.k.c.c.ConsumerConfig Thread-14-kafkaSpout-executor[4 4] [INFO] ConsumerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 partition.assignment.strategy = [org.apache.kafka.clients.consumer.RangeAssignor] reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 max.partition.fetch.bytes = 1048576 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS enable.auto.commit = false sasl.mechanism = GSSAPI interceptor.classes = null exclude.internal.topics = true ssl.truststore.password = null client.id = consumer-1 ssl.endpoint.identification.algorithm = null max.poll.records = 2147483647 check.crcs = true request.timeout.ms = 40000 heartbeat.interval.ms = 3000 auto.commit.interval.ms = 5000 receive.buffer.bytes = 65536 ssl.truststore.type = JKS ssl.truststore.location = null ssl.keystore.password = null fetch.min.bytes = 1 send.buffer.bytes = 131072 value.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer group.id = checkpoint_wifi_parser retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX ssl.key.password = null fetch.max.wait.ms = 500 sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 session.timeout.ms = 30000 metrics.num.samples = 2 key.deserializer = class org.apache.kafka.common.serialization.ByteArrayDeserializer ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 auto.offset.reset = latest 2018-05-25 13:49:24.734 o.a.k.c.p.ProducerConfig Thread-6-errorMessageWriter-executor[3 3] [WARN] The configuration request.required.acks = 1 was supplied but isn't a known config. 2018-05-25 13:49:24.738 o.a.k.c.u.AppInfoParser Thread-6-errorMessageWriter-executor[3 3] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:49:24.738 o.a.k.c.u.AppInfoParser Thread-6-errorMessageWriter-executor[3 3] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:49:24.741 o.a.s.d.executor Thread-6-errorMessageWriter-executor[3 3] [INFO] Prepared bolt errorMessageWriter:(3) 2018-05-25 13:49:24.787 o.a.k.c.u.AppInfoParser Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:49:24.787 o.a.k.c.u.AppInfoParser Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:49:24.789 o.a.s.k.s.NamedSubscription Thread-14-kafkaSpout-executor[4 4] [INFO] Kafka consumer subscribed topics [checkpoint_wifi] 2018-05-25 13:49:24.799 o.a.c.f.i.CuratorFrameworkImpl Thread-12-parserBolt-executor[5 5] [INFO] Starting 2018-05-25 13:49:24.819 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:zookeeper.version=3.4.6-235--1, built on 10/30/2017 02:17 GMT 2018-05-25 13:49:24.819 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:host.name=metron1.nic.in 2018-05-25 13:49:24.820 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.version=1.8.0_112 2018-05-25 13:49:24.820 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.vendor=Oracle Corporation 2018-05-25 13:49:24.820 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.home=/usr/jdk64/jdk1.8.0_112/jre 2018-05-25 13:49:24.820 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.class.path=/usr/hdp/2.6.3.0-235/storm/lib/asm-5.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-api-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-core-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.6.3.0-235/storm/lib/log4j-slf4j-impl-2.8.2.jar:/usr/hdp/2.6.3.0-235/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.6.3.0-235/storm/lib/objenesis-2.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.6.3.0-235/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/storm/lib/slf4j-api-1.7.21.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-core-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/storm-rename-hack-1.1.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/storm/lib/zookeeper.jar:/usr/hdp/2.6.3.0-235/storm/lib/ambari-metrics-storm-sink.jar:/usr/hdp/current/storm-supervisor/conf:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/stormjar.jar:/etc/hbase/conf:/etc/hadoop/conf 2018-05-25 13:49:24.820 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.library.path=/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources/Linux-amd64:/hadoop/storm/supervisor/stormdist/checkpoint_wifi-20-1527226482/resources:/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib 2018-05-25 13:49:24.821 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.io.tmpdir=/hadoop/storm/workers/d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a/tmp 2018-05-25 13:49:24.821 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:java.compiler= 2018-05-25 13:49:24.821 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.name=Linux 2018-05-25 13:49:24.821 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.arch=amd64 2018-05-25 13:49:24.821 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:os.version=3.10.0-693.17.1.el7.x86_64 2018-05-25 13:49:24.822 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.name=storm 2018-05-25 13:49:24.822 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.home=/home/storm 2018-05-25 13:49:24.822 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Client environment:user.dir=/hadoop/storm/workers/d16a3be8-a414-42d9-9e1c-f5ad5cbdc52a 2018-05-25 13:49:24.824 o.a.z.ZooKeeper Thread-12-parserBolt-executor[5 5] [INFO] Initiating client connection, connectString=master2.nic.in:2181,master1.nic.in:2181 sessionTimeout=60000 watcher=org.apache.curator.ConnectionState@29f5467a 2018-05-25 13:49:24.856 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Opening socket connection to server master1.nic.in/10.122.62.98:2181. Will not attempt to authenticate using SASL (unknown error) 2018-05-25 13:49:24.858 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Socket connection established, initiating session, client: /10.122.62.67:47990, server: master1.nic.in/10.122.62.98:2181 2018-05-25 13:49:24.869 o.a.z.ClientCnxn Thread-12-parserBolt-executor[5 5]-SendThread(master1.nic.in:2181) [INFO] Session establishment complete on server master1.nic.in/10.122.62.98:2181, sessionid = 0x163816421312c30, negotiated timeout = 60000 2018-05-25 13:49:24.882 o.a.c.f.s.ConnectionStateManager Thread-12-parserBolt-executor[5 5]-EventThread [INFO] State change: CONNECTED 2018-05-25 13:49:24.901 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Preparing Storm Metrics Sink 2018-05-25 13:49:24.918 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Topology name "checkpoint_wifi" contains "_" which can be problematic for AMS. 2018-05-25 13:49:24.918 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Encouraged to not using any of these strings: ".", "_" 2018-05-25 13:49:24.918 o.a.h.m.s.s.StormTimelineMetricsSink Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [WARN] Same suggestion applies to component name. 2018-05-25 13:49:24.920 o.a.s.d.executor Thread-4-__metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink-executor[2 2] [INFO] Prepared bolt __metricsorg.apache.hadoop.metrics2.sink.storm.StormTimelineMetricsSink:(2) 2018-05-25 13:49:25.001 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Discovered coordinator master1.nic.in:6667 (id: 2147482646 rack: null) for group checkpoint_wifi_parser. 2018-05-25 13:49:25.002 o.a.k.c.c.i.ConsumerCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Revoking previously assigned partitions [] for group checkpoint_wifi_parser 2018-05-25 13:49:25.002 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Partitions revoked. [consumer-group=checkpoint_wifi_parser, consumer=org.apache.kafka.clients.consumer.KafkaConsumer@19f6cc6d, topic-partitions=[]] 2018-05-25 13:49:25.003 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] (Re-)joining group checkpoint_wifi_parser 2018-05-25 13:49:25.600 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:49:25.609 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [INFO] ProducerConfig values: metric.reporters = [] metadata.max.age.ms = 300000 reconnect.backoff.ms = 50 sasl.kerberos.ticket.renew.window.factor = 0.8 bootstrap.servers = [master2.nic.in:6667, master1.nic.in:6667] ssl.keystore.type = JKS sasl.mechanism = GSSAPI max.block.ms = 60000 interceptor.classes = null ssl.truststore.password = null client.id = producer-2 ssl.endpoint.identification.algorithm = null request.timeout.ms = 30000 acks = 1 receive.buffer.bytes = 32768 ssl.truststore.type = JKS retries = 0 ssl.truststore.location = null ssl.keystore.password = null send.buffer.bytes = 131072 compression.type = none metadata.fetch.timeout.ms = 60000 retry.backoff.ms = 100 sasl.kerberos.kinit.cmd = /usr/bin/kinit buffer.memory = 33554432 timeout.ms = 30000 key.serializer = class org.apache.kafka.common.serialization.StringSerializer sasl.kerberos.service.name = null sasl.kerberos.ticket.renew.jitter = 0.05 ssl.trustmanager.algorithm = PKIX block.on.buffer.full = false ssl.key.password = null sasl.kerberos.min.time.before.relogin = 60000 connections.max.idle.ms = 540000 max.in.flight.requests.per.connection = 5 metrics.num.samples = 2 ssl.protocol = TLS ssl.provider = null ssl.enabled.protocols = [TLSv1.2, TLSv1.1, TLSv1] batch.size = 16384 ssl.keystore.location = null ssl.cipher.suites = null security.protocol = PLAINTEXT max.request.size = 1048576 value.serializer = class org.apache.kafka.common.serialization.StringSerializer ssl.keymanager.algorithm = SunX509 metrics.sample.window.ms = 30000 partitioner.class = class org.apache.kafka.clients.producer.internals.DefaultPartitioner linger.ms = 0 2018-05-25 13:49:25.610 o.a.k.c.p.ProducerConfig Thread-12-parserBolt-executor[5 5] [WARN] The configuration request.required.acks = 1 was supplied but isn't a known config. 2018-05-25 13:49:25.610 o.a.k.c.u.AppInfoParser Thread-12-parserBolt-executor[5 5] [INFO] Kafka version : 0.10.0.2.5.0.0-1245 2018-05-25 13:49:25.610 o.a.k.c.u.AppInfoParser Thread-12-parserBolt-executor[5 5] [INFO] Kafka commitId : dae559f56f07e2cd 2018-05-25 13:49:25.611 o.a.s.d.executor Thread-12-parserBolt-executor[5 5] [INFO] Prepared bolt parserBolt:(5) 2018-05-25 13:49:43.840 o.a.k.c.c.i.AbstractCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Successfully joined group checkpoint_wifi_parser with generation 10 2018-05-25 13:49:43.843 o.a.k.c.c.i.ConsumerCoordinator Thread-14-kafkaSpout-executor[4 4] [INFO] Setting newly assigned partitions [checkpoint_wifi-0] for group checkpoint_wifi_parser 2018-05-25 13:49:43.843 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Partitions reassignment. [consumer-group=checkpoint_wifi_parser, consumer=org.apache.kafka.clients.consumer.KafkaConsumer@19f6cc6d, topic-partitions=[checkpoint_wifi-0]] 2018-05-25 13:49:43.857 o.a.s.k.s.KafkaSpout Thread-14-kafkaSpout-executor[4 4] [INFO] Initialization complete 2018-05-25 13:51:45.271 o.a.h.m.s.t.a.MetricSinkWriteShardHostnameHashingStrategy Thread-19 [INFO] Calculated collector shard master2.nic.in based on hostname: metron1.nic.in 2018-05-25 13:59:13.363 o.a.s.d.executor Thread-14-kafkaSpout-executor[4 4] [INFO] Deactivating spout kafkaSpout:(4) 2018-05-25 13:59:45.115 o.a.s.d.worker Thread-17 [INFO] Shutting down worker checkpoint_wifi-20-1527226482 c2953577-5859-4402-96e9-1220aafde03a 6700 2018-05-25 13:59:45.116 o.a.s.d.worker Thread-17 [INFO] Terminating messaging context 2018-05-25 13:59:45.116 o.a.s.d.worker Thread-17 [INFO] Shutting down executors