Member since
12-14-2015
15
Posts
1
Kudos Received
0
Solutions
07-27-2016
11:23 PM
Good catch! I'm glad you took the time to write up your findings for others to learn.
... View more
12-17-2015
04:26 PM
Unfortunatly after the install it can't start the master with the following error. Traceback (most recent call last):
File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_master.py", line 24, in <module>
AccumuloScript('master').execute()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 219, in execute
method(env)
File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py", line 79, in start
self.configure(env) # for security
File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py", line 73, in configure
setup_conf_dir(name=self.component)
File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_configuration.py", line 167, in setup_conf_dir
mode=0700
File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 154, in __init__
self.env.run()
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
self.run_action(resource, action)
File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
provider_action()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 394, in action_create_on_execute
self.action_delayed("create")
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 391, in action_delayed
self.get_hdfs_resource_executor().action_delayed(action_name, self)
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 244, in action_delayed
self._assert_valid()
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 228, in _assert_valid
self.target_status = self._get_file_status(target)
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 284, in _get_file_status
list_status = self.util.run_command(target, 'GETFILESTATUS', method='GET', ignore_status_codes=['404'], assertable_result=False)
File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 202, in run_command
raise Fail(err_msg)
resource_management.core.exceptions.Fail: Execution of 'curl -sS -L -w '%{http_code}' -X GET 'http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503.
<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01//EN" "http://www.w3.org/TR/html4/strict.dtd">
<html><head>
<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
<title>ERROR: The requested URL could not be retrieved</title>
<style type="text/css"><!--
/*
Stylesheet for Squid Error pages
Adapted from design by Free CSS Templates
http://www.freecsstemplates.org
Released for free under a Creative Commons Attribution 2.5 License
*/
/* Page basics */
* {
font-family: verdana, sans-serif;
}
html body {
margin: 0;
padding: 0;
background: #efefef;
font-size: 12px;
color: #1e1e1e;
}
/* Page displayed title area */
#titles {
margin-left: 15px;
padding: 10px;
padding-left: 100px;
background: url('http://www.squid-cache.org/Artwork/SN.png') no-repeat left;
}
/* initial title */
#titles h1 {
color: #000000;
}
#titles h2 {
color: #000000;
}
/* special event: FTP success page titles */
#titles ftpsuccess {
background-color:#00ff00;
width:100%;
}
/* Page displayed body content area */
#content {
padding: 10px;
background: #ffffff;
}
/* General text */
p {
}
/* error brief description */
#error p {
}
/* some data which may have caused the problem */
#data {
}
/* the error message received from the system or other software */
#sysmsg {
}
pre {
font-family:sans-serif;
}
/* special event: FTP / Gopher directory listing */
#dirmsg {
font-family: courier;
color: black;
font-size: 10pt;
}
#dirlisting {
margin-left: 2%;
margin-right: 2%;
}
#dirlisting tr.entry td.icon,td.filename,td.size,td.date {
border-bottom: groove;
}
#dirlisting td.size {
width: 50px;
text-align: right;
padding-right: 5px;
}
/* horizontal lines */
hr {
margin: 0;
}
/* page displayed footer area */
#footer {
font-size: 9px;
padding-left: 10px;
}
body
:lang(fa) { direction: rtl; font-size: 100%; font-family: Tahoma, Roya, sans-serif; float: right; }
:lang(he) { direction: rtl; }
--></style>
</head><body id=ERR_DNS_FAIL>
<div id="titles">
<h1>ERROR</h1>
<h2>The requested URL could not be retrieved</h2>
</div>
<hr>
<div id="content">
<p>The following error was encountered while trying to retrieve the URL: <a href="http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?">http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?</a></p>
<blockquote id="error">
<p><b>Unable to determine IP address from host name <q>sandbox.hortonworks.com</q></b></p>
</blockquote>
<p>The DNS server returned:</p>
<blockquote id="data">
<pre>Name Error: The domain name does not exist.</pre>
</blockquote>
<p>This means that the cache was not able to resolve the hostname presented in the URL. Check if the address is correct.</p>
<p>Your cache administrator is <a href="mailto:hotline@kps-consulting.com?subject=CacheErrorInfo%20-%20ERR_DNS_FAIL&body=CacheHost%3A%20kg-ffm-proxy.kps.kps-consulting.com%0D%0AErrPage%3A%20ERR_DNS_FAIL%0D%0AErr%3A%20%5Bnone%5D%0D%0ADNS%20ErrMsg%3A%20Name%20Error%3A%20The%20domain%20name%20does%20not%20exist.%0D%0ATimeStamp%3A%20Thu,%2017%20Dec%202015%2016%3A16%3A52%20GMT%0D%0A%0D%0AClientIP%3A%2010.13.8.41%0D%0AServerIP%3A%20sandbox.hortonworks.com%0D%0A%0D%0AHTTP%20Request%3A%0D%0AGET%20%2Fwebhdfs%2Fv1%2Fuser%2Faccumulo%3Fop%3DGETFILESTATUS%26user.name%3Dhdfs%20HTTP%2F1.1%0AUser-Agent%3A%20curl%2F7.19.7%20(x86_64-redhat-linux-gnu)%20libcurl%2F7.19.7%20NSS%2F3.19.1%20Basic%20ECC%20zlib%2F1.2.3%20libidn%2F1.18%20libssh2%2F1.4.2%0D%0AHost%3A%20sandbox.hortonworks.com%3A50070%0D%0AAccept%3A%20*%2F*%0D%0AProxy-Connection%3A%20Keep-Alive%0D%0A%0D%0A%0D%0A">hotline@kps-consulting.com</a>.</p>
<br>
</div>
<hr>
<div id="footer">
<p>Generated Thu, 17 Dec 2015 16:16:52 GMT by kg-ffm-proxy.kps.kps-consulting.com (squid/3.1.19)</p>
<!-- ERR_DNS_FAIL -->
</div>
</body></html>
stdout: /var/lib/ambari-agent/data/output-417.txt
2015-12-17 16:16:52,049 - Group['hadoop'] {}
2015-12-17 16:16:52,050 - Group['users'] {}
2015-12-17 16:16:52,050 - Group['zeppelin'] {}
2015-12-17 16:16:52,050 - Group['knox'] {}
2015-12-17 16:16:52,051 - Group['ranger'] {}
2015-12-17 16:16:52,051 - Group['spark'] {}
2015-12-17 16:16:52,051 - User['oozie'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,051 - User['hive'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,052 - User['zeppelin'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,053 - User['ambari-qa'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,053 - User['flume'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,054 - User['hdfs'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,054 - User['knox'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,055 - User['ranger'] {'gid': 'hadoop', 'groups': ['ranger']}
2015-12-17 16:16:52,055 - User['storm'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,056 - User['spark'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,056 - User['mapred'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,057 - User['accumulo'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,058 - User['hbase'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,058 - User['tez'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,059 - User['zookeeper'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,059 - User['kafka'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,060 - User['falcon'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,060 - User['sqoop'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,060 - User['yarn'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,061 - User['hcat'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,061 - User['ams'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,062 - User['atlas'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,063 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-17 16:16:52,064 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2015-12-17 16:16:52,068 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2015-12-17 16:16:52,068 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
2015-12-17 16:16:52,069 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-17 16:16:52,070 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2015-12-17 16:16:52,074 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2015-12-17 16:16:52,074 - Group['hdfs'] {'ignore_failures': False}
2015-12-17 16:16:52,074 - User['hdfs'] {'ignore_failures': False, 'groups': ['hadoop', 'hdfs']}
2015-12-17 16:16:52,075 - Directory['/etc/hadoop'] {'mode': 0755}
2015-12-17 16:16:52,086 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2015-12-17 16:16:52,087 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0777}
2015-12-17 16:16:52,097 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'}
2015-12-17 16:16:52,102 - Skipping Execute[('setenforce', '0')] due to not_if
2015-12-17 16:16:52,102 - Directory['/var/log/hadoop'] {'owner': 'root', 'mode': 0775, 'group': 'hadoop', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,104 - Directory['/var/run/hadoop'] {'owner': 'root', 'group': 'root', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,105 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,109 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,110 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,111 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,118 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,118 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
2015-12-17 16:16:52,119 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'}
2015-12-17 16:16:52,123 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'}
2015-12-17 16:16:52,126 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755}
2015-12-17 16:16:52,244 - Directory['/usr/hdp/current/accumulo-master/conf'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True, 'mode': 0755}
2015-12-17 16:16:52,245 - Directory['/usr/hdp/current/accumulo-master/conf/server'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True, 'mode': 0700}
2015-12-17 16:16:52,246 - XmlConfig['accumulo-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/accumulo-master/conf/server', 'mode': 0600, 'configuration_attributes': {}, 'owner': 'accumulo', 'configurations': ...}
2015-12-17 16:16:52,255 - Generating config: /usr/hdp/current/accumulo-master/conf/server/accumulo-site.xml
2015-12-17 16:16:52,255 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-site.xml'] {'owner': 'accumulo', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0600, 'encoding': 'UTF-8'}
2015-12-17 16:16:52,266 - Directory['/var/run/accumulo'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True}
2015-12-17 16:16:52,267 - Directory['/var/log/accumulo'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True}
2015-12-17 16:16:52,270 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-env.sh'] {'content': InlineTemplate(...), 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,270 - PropertiesFile['/usr/hdp/current/accumulo-master/conf/server/client.conf'] {'owner': 'accumulo', 'group': 'hadoop', 'properties': {'instance.zookeeper.host': 'sandbox.hortonworks.com:2181', 'instance.name': 'hdp-accumulo-instance', 'instance.zookeeper.timeout': '30s'}}
2015-12-17 16:16:52,274 - Generating properties file: /usr/hdp/current/accumulo-master/conf/server/client.conf
2015-12-17 16:16:52,274 - File['/usr/hdp/current/accumulo-master/conf/server/client.conf'] {'owner': 'accumulo', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,276 - Writing File['/usr/hdp/current/accumulo-master/conf/server/client.conf'] because contents don't match
2015-12-17 16:16:52,276 - File['/usr/hdp/current/accumulo-master/conf/server/log4j.properties'] {'content': ..., 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,277 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/auditLog.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,279 - File['/usr/hdp/current/accumulo-master/conf/server/auditLog.xml'] {'content': Template('auditLog.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,279 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/generic_logger.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,282 - File['/usr/hdp/current/accumulo-master/conf/server/generic_logger.xml'] {'content': Template('generic_logger.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,283 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/monitor_logger.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,284 - File['/usr/hdp/current/accumulo-master/conf/server/monitor_logger.xml'] {'content': Template('monitor_logger.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,285 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-metrics.xml'] {'content': StaticFile('accumulo-metrics.xml'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,285 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/tracers'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,287 - File['/usr/hdp/current/accumulo-master/conf/server/tracers'] {'content': Template('tracers.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,287 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/gc'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,289 - File['/usr/hdp/current/accumulo-master/conf/server/gc'] {'content': Template('gc.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,289 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/monitor'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,291 - File['/usr/hdp/current/accumulo-master/conf/server/monitor'] {'content': Template('monitor.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,291 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/slaves'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,293 - File['/usr/hdp/current/accumulo-master/conf/server/slaves'] {'content': Template('slaves.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,293 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/masters'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,295 - File['/usr/hdp/current/accumulo-master/conf/server/masters'] {'content': Template('masters.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,295 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/hadoop-metrics2-accumulo.properties'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,299 - File['/usr/hdp/current/accumulo-master/conf/server/hadoop-metrics2-accumulo.properties'] {'content': Template('hadoop-metrics2-accumulo.properties.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,300 - HdfsResource['/user/accumulo'] {'security_enabled': False, 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'keytab': [EMPTY], 'default_fs': 'hdfs://sandbox.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': 'kinit', 'principal_name': [EMPTY], 'user': 'hdfs', 'owner': 'accumulo', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0700}
2015-12-17 16:16:52,302 - checked_call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmps_Hczm 2>/tmp/tmpLmfalF''] {'logoutput': None, 'quiet': False}
2015-12-17 16:16:52,325 - checked_call returned (0, '')
... View more
08-31-2016
02:09 PM
1 Kudo
Hi.
And is it possible to delete/remove these applications to free space on virtual machine ?
Thanks Valery.
... View more