Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Why does the install of Accumulo not work with HDP Sandbox?

Solved Go to solution
Highlighted

Why does the install of Accumulo not work with HDP Sandbox?

Explorer

I get the following error trying to install Accumulo on the HDP Sandbox. I need Accumulo to run sqoop.

Does anyone know how to fix this Accumulo installation error - seems like a repository error.

stderr: 
Traceback (most recent call last):
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py", line 65, in <module>
    AccumuloClient().execute()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 219, in execute
    method(env)
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_client.py", line 36, in install
    self.install_packages(env)
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 395, in install_packages
    Package(name)
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 154, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/__init__.py", line 45, in action_install
    self.install_package(package_name, self.resource.use_repos, self.resource.skip_repos)
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/package/yumrpm.py", line 49, in install_package
    shell.checked_call(cmd, sudo=True, logoutput=self.get_logoutput())
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 70, in inner
    result = function(command, **kwargs)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 92, in checked_call
    tries=tries, try_sleep=try_sleep)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 140, in _call_wrapper
    result = _call(command, **kwargs_copy)
  File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 291, in _call
    raise Fail(err_msg)
resource_management.core.exceptions.Fail: Execution of '/usr/bin/yum -d 0 -e 0 -y install 'accumulo_2_3_*'' returned 1. Error: Cannot find a valid baseurl for repo: base
Could not retrieve mirrorlist http://mirrorlist.centos.org/?release=6&arch=x86_... error was
14: PYCURL ERROR 6 - "Couldn't resolve host 'mirrorlist.centos.org'"










 stdout:
2015-12-15 15:04:12,132 - Group['hadoop'] {}
2015-12-15 15:04:12,133 - Group['users'] {}
2015-12-15 15:04:12,133 - Group['zeppelin'] {}
2015-12-15 15:04:12,134 - Group['knox'] {}
2015-12-15 15:04:12,134 - Group['ranger'] {}
2015-12-15 15:04:12,134 - Group['spark'] {}
2015-12-15 15:04:12,134 - User['oozie'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-15 15:04:12,135 - User['hive'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,135 - User['zeppelin'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,136 - User['ambari-qa'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-15 15:04:12,136 - User['flume'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,137 - User['hdfs'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,137 - User['knox'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,138 - User['ranger'] {'gid': 'hadoop', 'groups': ['ranger']}
2015-12-15 15:04:12,138 - User['storm'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,139 - User['spark'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,140 - User['mapred'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,140 - User['accumulo'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,140 - Adding user User['accumulo']
2015-12-15 15:04:12,227 - User['hbase'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,228 - User['tez'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-15 15:04:12,229 - User['zookeeper'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,229 - User['kafka'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,230 - User['falcon'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-15 15:04:12,230 - User['sqoop'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,231 - User['yarn'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,231 - User['hcat'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,232 - User['ams'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,232 - User['atlas'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-15 15:04:12,233 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-15 15:04:12,234 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2015-12-15 15:04:12,238 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2015-12-15 15:04:12,238 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
2015-12-15 15:04:12,239 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-15 15:04:12,240 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2015-12-15 15:04:12,243 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2015-12-15 15:04:12,244 - Group['hdfs'] {'ignore_failures': False}
2015-12-15 15:04:12,244 - User['hdfs'] {'ignore_failures': False, 'groups': ['hadoop', 'hdfs']}
2015-12-15 15:04:12,245 - Directory['/etc/hadoop'] {'mode': 0755}
2015-12-15 15:04:12,256 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2015-12-15 15:04:12,256 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0777}
2015-12-15 15:04:12,267 - Repository['HDP-2.3'] {'base_url': 'http://public-repo-1.hortonworks.com/HDP/centos6/2.x/updates/2.3.2.0/', 'action': ['create'], 'components': ['HDP', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP', 'mirror_list': None}
2015-12-15 15:04:12,319 - File['/etc/yum.repos.d/HDP.repo'] {'content': InlineTemplate(...)}
2015-12-15 15:04:12,332 - Repository['HDP-UTILS-1.1.0.20'] {'base_url': 'http://s3.amazonaws.com/dev.hortonworks.com/HDP-UTILS-1.1.0.20/repos/centos6', 'action': ['create'], 'components': ['HDP-UTILS', 'main'], 'repo_template': '[{{repo_id}}]\nname={{repo_id}}\n{% if mirror_list %}mirrorlist={{mirror_list}}{% else %}baseurl={{base_url}}{% endif %}\n\npath=/\nenabled=1\ngpgcheck=0', 'repo_file_name': 'HDP-UTILS', 'mirror_list': None}
2015-12-15 15:04:12,334 - File['/etc/yum.repos.d/HDP-UTILS.repo'] {'content': InlineTemplate(...)}
2015-12-15 15:04:12,346 - Package['unzip'] {}
2015-12-15 15:04:12,482 - Skipping installation of existing package unzip
2015-12-15 15:04:12,482 - Package['curl'] {}
2015-12-15 15:04:12,553 - Skipping installation of existing package curl
2015-12-15 15:04:12,553 - Package['hdp-select'] {}
2015-12-15 15:04:12,623 - Skipping installation of existing package hdp-select
2015-12-15 15:04:12,755 - Package['accumulo_2_3_*'] {}
2015-12-15 15:04:12,890 - Installing package accumulo_2_3_* ('/usr/bin/yum -d 0 -e 0 -y install 'accumulo_2_3_*'')

I think this might be the problem

14: PYCURL ERROR 6 - "Couldn't resolve host 'mirrorlist.centos.org'"
1 ACCEPTED SOLUTION

Accepted Solutions
Highlighted

Re: Why does the install of Accumulo not work with HDP Sandbox?

Contributor

It appears you cannot resolve mirrorlist.centos.org via DNS from your virtual machine.

Does the follow return a result?

nslookup mirrorlist.centos.org

If not, I expect you have configured the VM with a Host-Only adapter, which will not allow the VM to access the internet.

View solution in original post

4 REPLIES 4
Highlighted

Re: Why does the install of Accumulo not work with HDP Sandbox?

Contributor

It appears you cannot resolve mirrorlist.centos.org via DNS from your virtual machine.

Does the follow return a result?

nslookup mirrorlist.centos.org

If not, I expect you have configured the VM with a Host-Only adapter, which will not allow the VM to access the internet.

View solution in original post

Highlighted

Re: Why does the install of Accumulo not work with HDP Sandbox?

Explorer

Looks like the machine needed a proxy to connect to the internet. The proxy solved the problem. Now I have a bad situation as Ambari thinks accumulo is installed but it really isn't - I hope installing accumulo by hand will work out ...

yum -d 0 -e 0 -y install 'accumulo_2_3_*'

Re: Why does the install of Accumulo not work with HDP Sandbox?

Explorer

I just found the "reinstall" button in Ambari that seemed to help me after i broke the first installation. thanks for your help!

Highlighted

Re: Why does the install of Accumulo not work with HDP Sandbox?

Explorer

Unfortunatly after the install it can't start the master with the following error.

Traceback (most recent call last):
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_master.py", line 24, in <module>
    AccumuloScript('master').execute()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 219, in execute
    method(env)
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py", line 79, in start
    self.configure(env) # for security
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_script.py", line 73, in configure
    setup_conf_dir(name=self.component)
  File "/var/lib/ambari-agent/cache/common-services/ACCUMULO/1.6.1.2.2.0/package/scripts/accumulo_configuration.py", line 167, in setup_conf_dir
    mode=0700
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 154, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 152, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 118, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 394, in action_create_on_execute
    self.action_delayed("create")
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 391, in action_delayed
    self.get_hdfs_resource_executor().action_delayed(action_name, self)
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 244, in action_delayed
    self._assert_valid()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 228, in _assert_valid
    self.target_status = self._get_file_status(target)
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 284, in _get_file_status
    list_status = self.util.run_command(target, 'GETFILESTATUS', method='GET', ignore_status_codes=['404'], assertable_result=False)
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/providers/hdfs_resource.py", line 202, in run_command
    raise Fail(err_msg)
resource_management.core.exceptions.Fail: Execution of 'curl -sS -L -w '%{http_code}' -X GET 'http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. 
<!DOCTYPE html PUBLIC "-//W3C//DTD HTML 4.01//EN" "http://www.w3.org/TR/html4/strict.dtd">
<html><head>
<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
<title>ERROR: The requested URL could not be retrieved</title>
<style type="text/css"><!-- 
 /*
 Stylesheet for Squid Error pages
 Adapted from design by Free CSS Templates
 http://www.freecsstemplates.org
 Released for free under a Creative Commons Attribution 2.5 License
*/

/* Page basics */
* {
	font-family: verdana, sans-serif;
}

html body {
	margin: 0;
	padding: 0;
	background: #efefef;
	font-size: 12px;
	color: #1e1e1e;
}

/* Page displayed title area */
#titles {
	margin-left: 15px;
	padding: 10px;
	padding-left: 100px;
	background: url('http://www.squid-cache.org/Artwork/SN.png') no-repeat left;
}

/* initial title */
#titles h1 {
	color: #000000;
}
#titles h2 {
	color: #000000;
}

/* special event: FTP success page titles */
#titles ftpsuccess {
	background-color:#00ff00;
	width:100%;
}

/* Page displayed body content area */
#content {
	padding: 10px;
	background: #ffffff;
}

/* General text */
p {
}

/* error brief description */
#error p {
}

/* some data which may have caused the problem */
#data {
}

/* the error message received from the system or other software */
#sysmsg {
}

pre {
    font-family:sans-serif;
}

/* special event: FTP / Gopher directory listing */
#dirmsg {
    font-family: courier;
    color: black;
    font-size: 10pt;
}
#dirlisting {
    margin-left: 2%;
    margin-right: 2%;
}
#dirlisting tr.entry td.icon,td.filename,td.size,td.date {
    border-bottom: groove;
}
#dirlisting td.size {
    width: 50px;
    text-align: right;
    padding-right: 5px;
}

/* horizontal lines */
hr {
	margin: 0;
}

/* page displayed footer area */
#footer {
	font-size: 9px;
	padding-left: 10px;
}


body
:lang(fa) { direction: rtl; font-size: 100%; font-family: Tahoma, Roya, sans-serif; float: right; }
:lang(he) { direction: rtl; }
 --></style>
</head><body id=ERR_DNS_FAIL>
<div id="titles">
<h1>ERROR</h1>
<h2>The requested URL could not be retrieved</h2>
</div>
<hr>

<div id="content">
<p>The following error was encountered while trying to retrieve the URL: <a href="http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?">http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?</a></p>

<blockquote id="error">
<p><b>Unable to determine IP address from host name <q>sandbox.hortonworks.com</q></b></p>
</blockquote>

<p>The DNS server returned:</p>
<blockquote id="data">
<pre>Name Error: The domain name does not exist.</pre>
</blockquote>

<p>This means that the cache was not able to resolve the hostname presented in the URL. Check if the address is correct.</p>

<p>Your cache administrator is <a href="mailto:hotline@kps-consulting.com?subject=CacheErrorInfo%20-%20ERR_DNS_FAIL&body=CacheHost%3A%20kg-ffm-proxy.kps.kps-consulting.com%0D%0AErrPage%3A%20ERR_DNS_FAIL%0D%0AErr%3A%20%5Bnone%5D%0D%0ADNS%20ErrMsg%3A%20Name%20Error%3A%20The%20domain%20name%20does%20not%20exist.%0D%0ATimeStamp%3A%20Thu,%2017%20Dec%202015%2016%3A16%3A52%20GMT%0D%0A%0D%0AClientIP%3A%2010.13.8.41%0D%0AServerIP%3A%20sandbox.hortonworks.com%0D%0A%0D%0AHTTP%20Request%3A%0D%0AGET%20%2Fwebhdfs%2Fv1%2Fuser%2Faccumulo%3Fop%3DGETFILESTATUS%26user.name%3Dhdfs%20HTTP%2F1.1%0AUser-Agent%3A%20curl%2F7.19.7%20(x86_64-redhat-linux-gnu)%20libcurl%2F7.19.7%20NSS%2F3.19.1%20Basic%20ECC%20zlib%2F1.2.3%20libidn%2F1.18%20libssh2%2F1.4.2%0D%0AHost%3A%20sandbox.hortonworks.com%3A50070%0D%0AAccept%3A%20*%2F*%0D%0AProxy-Connection%3A%20Keep-Alive%0D%0A%0D%0A%0D%0A">hotline@kps-consulting.com</a>.</p>
<br>
</div>

<hr>
<div id="footer">
<p>Generated Thu, 17 Dec 2015 16:16:52 GMT by kg-ffm-proxy.kps.kps-consulting.com (squid/3.1.19)</p>
<!-- ERR_DNS_FAIL -->
</div>
</body></html>
            stdout:    /var/lib/ambari-agent/data/output-417.txt 

            2015-12-17 16:16:52,049 - Group['hadoop'] {}
2015-12-17 16:16:52,050 - Group['users'] {}
2015-12-17 16:16:52,050 - Group['zeppelin'] {}
2015-12-17 16:16:52,050 - Group['knox'] {}
2015-12-17 16:16:52,051 - Group['ranger'] {}
2015-12-17 16:16:52,051 - Group['spark'] {}
2015-12-17 16:16:52,051 - User['oozie'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,051 - User['hive'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,052 - User['zeppelin'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,053 - User['ambari-qa'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,053 - User['flume'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,054 - User['hdfs'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,054 - User['knox'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,055 - User['ranger'] {'gid': 'hadoop', 'groups': ['ranger']}
2015-12-17 16:16:52,055 - User['storm'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,056 - User['spark'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,056 - User['mapred'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,057 - User['accumulo'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,058 - User['hbase'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,058 - User['tez'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,059 - User['zookeeper'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,059 - User['kafka'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,060 - User['falcon'] {'gid': 'hadoop', 'groups': ['users']}
2015-12-17 16:16:52,060 - User['sqoop'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,060 - User['yarn'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,061 - User['hcat'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,061 - User['ams'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,062 - User['atlas'] {'gid': 'hadoop', 'groups': ['hadoop']}
2015-12-17 16:16:52,063 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-17 16:16:52,064 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2015-12-17 16:16:52,068 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2015-12-17 16:16:52,068 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'recursive': True, 'mode': 0775, 'cd_access': 'a'}
2015-12-17 16:16:52,069 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2015-12-17 16:16:52,070 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2015-12-17 16:16:52,074 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2015-12-17 16:16:52,074 - Group['hdfs'] {'ignore_failures': False}
2015-12-17 16:16:52,074 - User['hdfs'] {'ignore_failures': False, 'groups': ['hadoop', 'hdfs']}
2015-12-17 16:16:52,075 - Directory['/etc/hadoop'] {'mode': 0755}
2015-12-17 16:16:52,086 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2015-12-17 16:16:52,087 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0777}
2015-12-17 16:16:52,097 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'}
2015-12-17 16:16:52,102 - Skipping Execute[('setenforce', '0')] due to not_if
2015-12-17 16:16:52,102 - Directory['/var/log/hadoop'] {'owner': 'root', 'mode': 0775, 'group': 'hadoop', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,104 - Directory['/var/run/hadoop'] {'owner': 'root', 'group': 'root', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,105 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'recursive': True, 'cd_access': 'a'}
2015-12-17 16:16:52,109 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,110 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,111 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,118 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs'}
2015-12-17 16:16:52,118 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
2015-12-17 16:16:52,119 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'}
2015-12-17 16:16:52,123 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'}
2015-12-17 16:16:52,126 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755}
2015-12-17 16:16:52,244 - Directory['/usr/hdp/current/accumulo-master/conf'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True, 'mode': 0755}
2015-12-17 16:16:52,245 - Directory['/usr/hdp/current/accumulo-master/conf/server'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True, 'mode': 0700}
2015-12-17 16:16:52,246 - XmlConfig['accumulo-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/accumulo-master/conf/server', 'mode': 0600, 'configuration_attributes': {}, 'owner': 'accumulo', 'configurations': ...}
2015-12-17 16:16:52,255 - Generating config: /usr/hdp/current/accumulo-master/conf/server/accumulo-site.xml
2015-12-17 16:16:52,255 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-site.xml'] {'owner': 'accumulo', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0600, 'encoding': 'UTF-8'}
2015-12-17 16:16:52,266 - Directory['/var/run/accumulo'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True}
2015-12-17 16:16:52,267 - Directory['/var/log/accumulo'] {'owner': 'accumulo', 'group': 'hadoop', 'recursive': True}
2015-12-17 16:16:52,270 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-env.sh'] {'content': InlineTemplate(...), 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,270 - PropertiesFile['/usr/hdp/current/accumulo-master/conf/server/client.conf'] {'owner': 'accumulo', 'group': 'hadoop', 'properties': {'instance.zookeeper.host': 'sandbox.hortonworks.com:2181', 'instance.name': 'hdp-accumulo-instance', 'instance.zookeeper.timeout': '30s'}}
2015-12-17 16:16:52,274 - Generating properties file: /usr/hdp/current/accumulo-master/conf/server/client.conf
2015-12-17 16:16:52,274 - File['/usr/hdp/current/accumulo-master/conf/server/client.conf'] {'owner': 'accumulo', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,276 - Writing File['/usr/hdp/current/accumulo-master/conf/server/client.conf'] because contents don't match
2015-12-17 16:16:52,276 - File['/usr/hdp/current/accumulo-master/conf/server/log4j.properties'] {'content': ..., 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,277 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/auditLog.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,279 - File['/usr/hdp/current/accumulo-master/conf/server/auditLog.xml'] {'content': Template('auditLog.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,279 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/generic_logger.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,282 - File['/usr/hdp/current/accumulo-master/conf/server/generic_logger.xml'] {'content': Template('generic_logger.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,283 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/monitor_logger.xml'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,284 - File['/usr/hdp/current/accumulo-master/conf/server/monitor_logger.xml'] {'content': Template('monitor_logger.xml.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,285 - File['/usr/hdp/current/accumulo-master/conf/server/accumulo-metrics.xml'] {'content': StaticFile('accumulo-metrics.xml'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': 0644}
2015-12-17 16:16:52,285 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/tracers'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,287 - File['/usr/hdp/current/accumulo-master/conf/server/tracers'] {'content': Template('tracers.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,287 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/gc'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,289 - File['/usr/hdp/current/accumulo-master/conf/server/gc'] {'content': Template('gc.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,289 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/monitor'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,291 - File['/usr/hdp/current/accumulo-master/conf/server/monitor'] {'content': Template('monitor.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,291 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/slaves'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,293 - File['/usr/hdp/current/accumulo-master/conf/server/slaves'] {'content': Template('slaves.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,293 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/masters'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,295 - File['/usr/hdp/current/accumulo-master/conf/server/masters'] {'content': Template('masters.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,295 - TemplateConfig['/usr/hdp/current/accumulo-master/conf/server/hadoop-metrics2-accumulo.properties'] {'owner': 'accumulo', 'template_tag': None, 'group': 'hadoop'}
2015-12-17 16:16:52,299 - File['/usr/hdp/current/accumulo-master/conf/server/hadoop-metrics2-accumulo.properties'] {'content': Template('hadoop-metrics2-accumulo.properties.j2'), 'owner': 'accumulo', 'group': 'hadoop', 'mode': None}
2015-12-17 16:16:52,300 - HdfsResource['/user/accumulo'] {'security_enabled': False, 'hadoop_bin_dir': '/usr/hdp/current/hadoop-client/bin', 'keytab': [EMPTY], 'default_fs': 'hdfs://sandbox.hortonworks.com:8020', 'hdfs_site': ..., 'kinit_path_local': 'kinit', 'principal_name': [EMPTY], 'user': 'hdfs', 'owner': 'accumulo', 'hadoop_conf_dir': '/usr/hdp/current/hadoop-client/conf', 'type': 'directory', 'action': ['create_on_execute'], 'mode': 0700}
2015-12-17 16:16:52,302 - checked_call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET '"'"'http://sandbox.hortonworks.com:50070/webhdfs/v1/user/accumulo?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmps_Hczm 2>/tmp/tmpLmfalF''] {'logoutput': None, 'quiet': False}
2015-12-17 16:16:52,325 - checked_call returned (0, '')
Don't have an account?
Coming from Hortonworks? Activate your account here