stderr: Traceback (most recent call last): File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_master.py", line 93, in ParserMaster().execute() File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 280, in execute method(env) File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_master.py", line 81, in restart commands.restart_parser_topologies(env) File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_commands.py", line 146, in restart_parser_topologies self.stop_parser_topologies() File "/var/lib/ambari-agent/cache/common-services/METRON/0.4.0/package/scripts/parser_commands.py", line 141, in stop_parser_topologies Execute(stop_cmd, user=self.__params.metron_user) File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 155, in __init__ self.env.run() File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 160, in run self.run_action(resource, action) File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 124, in run_action provider_action() File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 273, in action_run tries=self.resource.tries, try_sleep=self.resource.try_sleep) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 70, in inner result = function(command, **kwargs) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 92, in checked_call tries=tries, try_sleep=try_sleep) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 140, in _call_wrapper result = _call(command, **kwargs_copy) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 293, in _call raise ExecutionFailed(err_msg, code, out, err) resource_management.core.exceptions.ExecutionFailed: Execution of 'storm kill bro' returned 1. Running: /usr/jdk64/jdk1.8.0_77/bin/java -client -Ddaemon.name= -Dstorm.options= -Dstorm.home=/usr/hdp/2.5.3.0-37/storm -Dstorm.log.dir=/var/log/storm -Djava.library.path=/usr/local/lib:/opt/local/lib:/usr/lib:/usr/hdp/current/storm-client/lib -Dstorm.conf.file= -cp /usr/hdp/2.5.3.0-37/storm/lib/log4j-core-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/minlog-1.3.0.jar:/usr/hdp/2.5.3.0-37/storm/lib/objenesis-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/ring-cors-0.1.5.jar:/usr/hdp/2.5.3.0-37/storm/lib/storm-core-1.0.1.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/lib/disruptor-3.3.2.jar:/usr/hdp/2.5.3.0-37/storm/lib/kryo-3.0.3.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-over-slf4j-1.6.6.jar:/usr/hdp/2.5.3.0-37/storm/lib/reflectasm-1.10.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-slf4j-impl-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/log4j-api-2.1.jar:/usr/hdp/2.5.3.0-37/storm/lib/clojure-1.7.0.jar:/usr/hdp/2.5.3.0-37/storm/lib/zookeeper.jar:/usr/hdp/2.5.3.0-37/storm/lib/servlet-api-2.5.jar:/usr/hdp/2.5.3.0-37/storm/lib/storm-rename-hack-1.0.1.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/lib/slf4j-api-1.7.7.jar:/usr/hdp/2.5.3.0-37/storm/lib/asm-5.0.3.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ojdbc6.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ranger-storm-plugin-shim-0.6.0.2.5.3.0-37.jar:/usr/hdp/2.5.3.0-37/storm/extlib-daemon/ranger-plugin-classloader-0.6.0.2.5.3.0-37.jar:/usr/hdp/current/storm-supervisor/conf:/usr/hdp/2.5.3.0-37/storm/bin org.apache.storm.command.kill_topology bro Exception in thread "main" org.apache.storm.utils.NimbusLeaderNotFoundException: Found nimbuses [] none of which is elected as leader, please try again after some time. at org.apache.storm.utils.NimbusClient.getConfiguredClientAs(NimbusClient.java:84) at org.apache.storm.command.kill_topology$_main.doInvoke(kill_topology.clj:26) at clojure.lang.RestFn.applyTo(RestFn.java:137) at org.apache.storm.command.kill_topology.main(Unknown Source) stdout: 2017-06-09 21:52:14,307 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.3.0-37 2017-06-09 21:52:14,308 - Checking if need to create versioned conf dir /etc/hadoop/2.5.3.0-37/0 2017-06-09 21:52:14,308 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-06-09 21:52:14,329 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist already', '') 2017-06-09 21:52:14,329 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-06-09 21:52:14,349 - checked_call returned (0, '') 2017-06-09 21:52:14,349 - Ensuring that hadoop has the correct symlink structure 2017-06-09 21:52:14,350 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-06-09 21:52:14,454 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.3.0-37 2017-06-09 21:52:14,454 - Checking if need to create versioned conf dir /etc/hadoop/2.5.3.0-37/0 2017-06-09 21:52:14,455 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-06-09 21:52:14,475 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist already', '') 2017-06-09 21:52:14,476 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-06-09 21:52:14,495 - checked_call returned (0, '') 2017-06-09 21:52:14,496 - Ensuring that hadoop has the correct symlink structure 2017-06-09 21:52:14,496 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-06-09 21:52:14,497 - Group['metron'] {} 2017-06-09 21:52:14,498 - Group['livy'] {} 2017-06-09 21:52:14,499 - Group['elasticsearch'] {} 2017-06-09 21:52:14,499 - Group['spark'] {} 2017-06-09 21:52:14,499 - Group['hadoop'] {} 2017-06-09 21:52:14,499 - Group['kibana'] {} 2017-06-09 21:52:14,499 - Group['users'] {} 2017-06-09 21:52:14,500 - User['storm'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,500 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,501 - User['metron'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,501 - User['livy'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,502 - User['elasticsearch'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,502 - User['spark'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,503 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']} 2017-06-09 21:52:14,503 - User['kafka'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,504 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,504 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,505 - User['kibana'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,505 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,506 - User['hbase'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-06-09 21:52:14,506 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2017-06-09 21:52:14,508 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'} 2017-06-09 21:52:14,512 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if 2017-06-09 21:52:14,512 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'} 2017-06-09 21:52:14,513 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2017-06-09 21:52:14,514 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'} 2017-06-09 21:52:14,518 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if 2017-06-09 21:52:14,518 - Group['hdfs'] {} 2017-06-09 21:52:14,518 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'hdfs']} 2017-06-09 21:52:14,519 - FS Type: 2017-06-09 21:52:14,519 - Directory['/etc/hadoop'] {'mode': 0755} 2017-06-09 21:52:14,534 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'} 2017-06-09 21:52:14,534 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777} 2017-06-09 21:52:14,547 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'} 2017-06-09 21:52:14,560 - Directory['/var/log/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'} 2017-06-09 21:52:14,561 - Directory['/var/run/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'root', 'cd_access': 'a'} 2017-06-09 21:52:14,561 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'create_parents': True, 'cd_access': 'a'} 2017-06-09 21:52:14,565 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'} 2017-06-09 21:52:14,567 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'} 2017-06-09 21:52:14,568 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} 2017-06-09 21:52:14,581 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs', 'group': 'hadoop'} 2017-06-09 21:52:14,581 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755} 2017-06-09 21:52:14,582 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'} 2017-06-09 21:52:14,586 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'} 2017-06-09 21:52:14,590 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755} 2017-06-09 21:52:14,778 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.3.0-37 2017-06-09 21:52:14,778 - Checking if need to create versioned conf dir /etc/hadoop/2.5.3.0-37/0 2017-06-09 21:52:14,778 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-06-09 21:52:14,798 - call returned (1, '/etc/hadoop/2.5.3.0-37/0 exist already', '') 2017-06-09 21:52:14,799 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.3.0-37', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-06-09 21:52:14,818 - checked_call returned (0, '') 2017-06-09 21:52:14,818 - Ensuring that hadoop has the correct symlink structure 2017-06-09 21:52:14,818 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-06-09 21:52:14,820 - Create Metron Local Config Directory 2017-06-09 21:52:14,820 - Configure Metron global.json 2017-06-09 21:52:14,821 - Directory['/usr/metron/0.4.0/config/zookeeper'] {'owner': 'metron', 'group': 'metron', 'mode': 0755} 2017-06-09 21:52:14,826 - File['/usr/metron/0.4.0/config/zookeeper/global.json'] {'content': InlineTemplate(...), 'owner': 'metron'} 2017-06-09 21:52:14,828 - File['/usr/metron/0.4.0/config/zookeeper/../elasticsearch.properties'] {'content': InlineTemplate(...), 'owner': 'metron'} 2017-06-09 21:52:14,828 - Writing File['/usr/metron/0.4.0/config/zookeeper/../elasticsearch.properties'] because contents don't match 2017-06-09 21:52:14,829 - Loading config into ZooKeeper 2017-06-09 21:52:14,829 - Execute['/usr/metron/0.4.0/bin/zk_load_configs.sh --mode PUSH -i /usr/metron/0.4.0/config/zookeeper -z ec2-34-209-246-58.us-west-2.compute.amazonaws.com:2181'] {'path': ['/usr/jdk64/jdk1.8.0_77/bin']} 2017-06-09 21:52:16,440 - Calling security setup 2017-06-09 21:52:16,440 - Restarting the parser topologies 2017-06-09 21:52:16,440 - Stopping parsers 2017-06-09 21:52:16,440 - Stopping bro 2017-06-09 21:52:16,440 - Execute['storm kill bro'] {'user': 'metron'} Command failed after 1 tries